Unnamed: 0
int64 0
7.24k
| id
int64 1
7.28k
| raw_text
stringlengths 9
124k
| vw_text
stringlengths 12
15k
|
---|---|---|---|
1,200 | 2,094 | Kernel Feature Spaces and
Nonlinear Blind Source Separation
Stefan Harmeling1?, Andreas Ziehe1 , Motoaki Kawanabe1, Klaus-Robert M?ller1,2
1
Fraunhofer FIRST.IDA, Kekul?str. 7, 12489 Berlin, Germany
2
University of Potsdam, Department of Computer Science,
August-Bebel-Strasse 89, 14482 Potsdam, Germany
{harmeli,ziehe,kawanabe,klaus}@first.fhg.de
Abstract
In kernel based learning the data is mapped to a kernel feature space of
a dimension that corresponds to the number of training data points. In
practice, however, the data forms a smaller submanifold in feature space,
a fact that has been used e.g. by reduced set techniques for SVMs. We
propose a new mathematical construction that permits to adapt to the intrinsic dimension and to find an orthonormal basis of this submanifold.
In doing so, computations get much simpler and more important our
theoretical framework allows to derive elegant kernelized blind source
separation (BSS) algorithms for arbitrary invertible nonlinear mixings.
Experiments demonstrate the good performance and high computational
efficiency of our kTDSEP algorithm for the problem of nonlinear BSS.
1 Introduction
In a widespread area of applications kernel based learning machines, e.g. Support Vector
Machines (e.g. [19, 6]) give excellent solutions. This holds both for problems of supervised
and unsupervised learning (e.g. [3, 16, 12]). The general idea is to map the data x i (i =
1, . . . , T ) into some kernel feature space F by some mapping ? : <n ? F. Performing
a simple linear algorithm in F, then corresponds to a nonlinear algorithm in input space.
Essential ingredients to kernel based learning are (a) VC theory that can provide a relation
between the complexity of the function class in use and the generalization error and (b) the
famous kernel trick
k(x, y) = ?(x) ? ?(y),
(1)
which allows to efficiently compute scalar products. This trick is essential if e.g. F is
an infinite dimensional space. Note that even though F might be infinite dimensional the
subspace where the data lies is maximally T -dimensional. However, the data typically
forms an even smaller subspace in F (cf. also reduced set methods [15]). In this work
we therefore propose a new mathematical construction that allows us to adapt to the intrinsic dimension and to provide an orthonormal basis of this submanifold. Furthermore,
this makes computations much simpler and provides the basis for a new set of kernelized
learning algorithms.
?
To whom correspondence should be addressed.
To demonstrate the power of our new framework we will focus on the problem of nonlinear
BSS [2, 18, 9, 10, 20, 11, 13, 14, 7, 17, 8] and provide an elegant kernel based algorithm
for arbitrary invertible nonlinearities. In nonlinear BSS we observe a mixed signal of the
following structure
xt = f (st ),
(2)
where xt and st are n ? 1 column vectors and f is a possibly nonlinear function from < n to
<n . In the special case where f is an n?n matrix we retrieve standard linear BSS (e.g. [8, 4]
and references therein). Nonlinear BSS has so far been only applied to industrial pulp data
[8], but a large class of applications where nonlinearities can occur in the mixing process
are conceivable, e.g. in the fields of telecommunications, array processing, biomedical data
analysis (EEG, MEG, EMG, . . .) and acoustic source separation. Most research has so far
centered on post-nonlinear models, i.e.
xt = f (Ast ),
(3)
where A is a linear mixing matrix and f is a post-nonlinearity that operates componentwise.
Algorithmic solutions of eq.(3) have used e.g. self-organizing maps [13, 10], extensions of
GTM [14], neural networks [2, 11] or ensemble learning [18] to unfold the nonlinearity
f . Also a kernel based method was tried on very simple toy signals; however with some
stability problems [7]. Note, that all existing methods are of high computational cost and
depending on the algorithm are prone to run into local minima. In our contribution to the
general invertable nonlinear BSS case we apply a standard BSS technique [21, 1] (that
relies on temporal correlations) to mapped signals in feature space (cf. section 3). This is
not only mathematically elegant (cf. section 2), but proves to be a remarkably stable and
efficient algorithm with high performance, as we will see in the experiments on nonlinear
mixtures of toy and speech data (cf. section 4). Finally, a conclusion is given in section 5.
2 Theory
An orthonormal basis for a subspace in F
In order to establish a linear problem in feature space that corresponds to some nonlinear problem in input space we need to specify how to map inputs x 1 , . . . , xT ? <n into
the feature space F and how to handle its possibly high dimensionality. In addition to
the inputs, consider some further points v1 , . . . , vd ? <n from the same space, that will
later generate a basis in F. Alternatively, we could use kernel PCA [16]. However, in
this paper we concentrate on a different method. Let us denote the mapped points by
?x := [?(x1 ) ? ? ? ?(xT )] and ?v := [?(v1 ) ? ? ? ?(vd )]. We assume that the columns of
?v constitute a basis of the column space1 of ?x , which we note by
span(?v ) = span(?x ) and rank(?v ) = d.
(4)
Moreover, ?v being a basis implies that the matrix ?>
v ?v has full rank and its inverse
exists. So, now we can define an orthonormal basis
1
?2
? := ?v (?>
v ?v )
(5)
the column space of which is identical to the column space of ?v . Consequently this basis
? enables us to parameterize all vectors that lie in the column space of ? x by some vectors
PT
in <d . For instance for vectors i=1 ??i ?(xi ), which we write more compactly as ?x ?? ,
and ?x ?? in the column space of ?x with ?? and ?? in <T there exist ?? and ?? in <d
such that ?x ?? = ??? and ?x ?? = ??? . The orthonormality implies
>
> >
>
?>
? ?x ?x ?? = ?? ? ??? = ?? ??
(6)
input space
<
n
feature space
span(?)
F
parameter space
<d
Figure 1: Input data are mapped to some submanifold of F which is in the span of some ddimensional orthonormal basis ?. Therefore these mapped points can be parametrized in <d . The
linear directions in parameter space correspond to nonlinear directions in input space.
which states the remarkable property that the dot product of two linear combinations of the
columns of ?x in F coincides with the dot product in <d . By construction of ? (cf. eq.(5))
the column space of ?x is naturally isomorphic (as vector spaces) to <d . Moreover, this
isomorphism is compatible with the two involved dot products as was shown in eq.(6). This
implies that all properties regarding angles and lengths can be taken back and forth between
the column space of ?x and <d . The space that is spanned by ? is called parameter space.
Figure 1 pictures our intuition: Usually kernel methods parameterize the column space of
?x in terms of the mapped patterns {?(xi )} which effectively corresponds to vectors in
<T . The orthonormal basis from eq.(5), however enables us to work in < d i.e. in the span
of ?, which is extremely valuable since d depends solely on the kernel function and the
dimensionality of the input space. So d is independent of T .
Mapping inputs
Having established the machinery above, we will now show how to map the input data to
the right space. The expressions
>
(?>
v ?v )ij = ?(vi ) ?(vj ) = k(vi , vj )
with i, j = 1 . . . d
are the entries of a real valued d ? d matrix ?>
v ?v that can be effectively calculated using
the kernel trick and by construction of v1 , . . . , vd , it has full rank and is thus invertible.
Similarly we get
>
(?>
v ?x )ij = ?(vi ) ?(xj ) = k(vi , xj ) with i = 1 . . . d,
j = 1...T,
which are the entries of the real valued d ? T matrix ?>
v ?x . Using both matrices we
compute finally the parameter matrix
1
?2 >
?v ?x
?x := ?> ?x = (?>
v ?v )
(7)
1
The column space of ?x is the space that is spanned by the column vectors of ?x , written
span(?x ).
1
?2
which is also a real valued d ? T matrix; note that (?>
is symmetric. Regarding
v ?v )
computational costs, we have to evaluate the kernel function O(d 2 ) + O(dT ) times and
eq.(7) requires O(d3 ) multiplications; again note that d is much smaller than T . Furthermore storage requirements are cheaper as we do not have to hold the full T ? T kernel
matrix but only a d ? T matrix. Also, kernel based algorithms often require centering in
F, which in our setting is equivalent to centering in <d . Fortunately the latter can be done
quite cheaply.
Choosing vectors for the basis in F
So far we have assumed to have points v1 , . . . , vd that fulfill eq.(4) and we presented
the beneficial properties of our construction. In fact, v1 , . . . , vd are roughly analogous
to a reduced set in the support vector world [15]. Note however that often we can only
approximately fulfill eq.(4), i.e.
span(?v ) ? span(?x ).
(8)
In this case we strive for points that provide the best approximation.
Obviously d is finite since it is bounded by T , the number of inputs, and by the dimensionality of the feature space. Before formulating the algorithm we define the function rk(n)
for numbers n by the following process: randomly pick n points v 1 , . . . , vn from the inputs
and compute the rank of the corresponding n ? n matrix ?>
v ?v . Repeating this random
sampling process several times (e.g. 100 times) stabilizes this process in practice. Then we
denote by rk(n) the largest achieved rank; note that rk(n) ? n. Using this definition we
can formulate a recipe to find d (the dimension of the subspace of F): (1) start with a large
d with rk(d) < d. (2) Decrement d by one as long as rk(d) < d holds. As soon as we
have rk(d) = d we found the d. Choose v1 , . . . , vd as the vectors that achieve rank d. As
an alternative to random sampling we have also employed k-means clustering with similar
results.
3 Nonlinear blind source separation
To demonstrate the use of the orthonormal basis in F, we formulate a new nonlinear BSS
algorithm based on TDSEP [21]. We start from a set of points v1 , . . . , vd , that are provided
by the algorithm from the last section such that eq.(4) holds. Next, we use eq.(7) to compute
1
?2 >
?x [t] := ?> ?(x[t]) = (?>
?v ?(x[t])
v ?v )
? <d .
Hereby we have transformed the time signals x[t] from input space to parameter space signals ?x [t] (cf. Fig.1). Now we apply the usual TDSEP ([21]) that relies on simultaneous
diagonalisation techniques [5] to perform linear blind source separation on ? x [t] to obtain
d linear directions of separated nonlinear components in input space. This new algorithm is
denoted as kTDSEP (kernel TDSEP); in short, kTDSEP is TDSEP on the parameter space
defined in Fig.1. A key to the success of our algorithm are the time correlations exploited
by TDSEP; intuitively they provide the ?glue? that yields the coherence for the separated
signals. Note that for a linear kernel functions the new algorithm performs linear BSS.
Therefore linear BSS can be seen as a special case of our algorithm.
Note that common kernel based algorithms which do not use the d-dimensional orthonormal basis will run into computational problems. They need to hold and compute with a
kernel matrix that is T ? T instead of d ? T with T d in BSS problems. A further
problem is that manipulating such a T ? T matrix can easily become unstable. Moreover
BSS methods typically become unfeasible for separation problems of dimension T .
!
Figure 2: Scatterplot of x1 vs x2 for nonlinear mixing and demixing (upper left and right) and linear
demixing and true source signals (lower left and right). Note, that the nonlinear unmixing agrees very
nicely with the scatterplot of the true source signal.
4 Experiments
In the first experiment the source signals s[t] = [s1 [t] s2 [t]]> are a sinusoidal and a sawtooth signal with 2000 samples each. The nonlinearly mixed signals are defined as (cf. Fig.2
upper left panel)
x1 [t]
x2 [t]
= exp(s1 [t]) ? exp(s2 [t])
= exp(?s1 [t]) + exp(?s2 [t]).
A dimension d = 22 of the manifold in feature space was obtained by kTDSEP using
a polynomial kernel k(x, y) = (x> y + 1)6 by sampling from the inputs. The basisgenerating vectors v1 , . . . , v22 are shown as big dots in the upper left panel of Figure
2. Applying TDSEP to the 22 dimensional mapped signals ?x [t] we get 22 components
in parameter space. A scatter plot with the two components that best match the source
signals are shown in the right upper panel of Figure 2. The left lower panel also shows for
comparison the two components that we obtained by applying linear TDSEP directly to the
mixed signals x[t]. The plots clearly indicate that kTDSEP has unfolded the nonlinearity
successfully while the linear demixing algorithm failed.
In a second experiment two speech signals (with 20000 samples, sampling rate 8 kHz) that
are nonlinearly mixed by
x1 [t]
= s1 [t] + s32 [t]
x2 [t]
= s31 [t] + tanh(s2 [t]).
This time we used a Gaussian RBF kernel k(x, y) = exp(?|x ? y|2 ). kTDSEP identified
d = 41 and used k-means clustering to obtain v1 , . . . , v41 . These points are marked as
?+? in the left panel of figure 4. An application of TDSEP to the 41 dimensional parameter
mixture
s1
s2
kTDSEP
TDSEP
x1
x2
u1
u2
u1
u2
0.56
0.63
0.72
0.46
0.89
0.04
0.07
0.86
0.09
0.31
0.72
0.55
Table 3: Correlation coefficients for the signals shown in Fig.4.
space yields nonlinear components whose projections to the input space are depicted in the
right lower panel. We can see that linear TDSEP (right middle panel) failed and that the
directions of best matching kTDSEP components closely resemble the sources.
To confirm this visual impression we calculated the correlation coefficients of the kTDSEP
and TDSEP solution to the source signals (cf. table 3). Clearly, kTDSEP outperforms the
linear TDSEP algorithm, which is of course what one expects.
5 Conclusion
Our work has two main contributions. First, we propose a new formulation in the field of
kernel based learning methods that allows to construct an orthonormal basis of the subspace
of kernel feature space F where the data lies. This technique establishes a highly useful
(scalar product preserving) isomorphism between the image of the data points in F and a
d-dimensional space <d . Several interesting things follow: we can construct a new set of
efficient kernel-based algorithms e.g. a new and eventually more stable variant of kernel
PCA [16]. Moreover, we can acquire knowledge about the intrinsic dimension of the data
manifold in F from the learning process.
Second, using our new formulation we tackle the problem of nonlinear BSS from the viewpoint of kernel based learning. The proposed kTDSEP algorithm allows to unmix arbitrary
invertible nonlinear mixtures with low computational costs. Note, that the important ingredients are the temporal correlations of the source signals used by TDSEP. Experiments on
toy and speech signals underline that an elegant solution has been found to a challenging
problem.
Applications where nonlinearly mixed signals can occur, are found e.g. in the fields of
telecommunications, array processing, biomedical data analysis (EEG, MEG, EMG, . . .)
and acoustic source separation. In fact, our algorithm would allow to provide a softwarebased correction of sensors that have a nonlinear characteristics e.g. due to manufacturing
errors. Clearly kTDSEP is only one algorithm that can perform nonlinear BSS; kernelizing
other ICA algorithms can be done following our reasoning.
Acknowledgements The authors thank Benjamin Blankertz, Gunnar R?tsch, Sebastian
Mika for valuable discussions. This work was partly supported by the EU project (IST1999-14190 ? BLISS) and DFG (JA 379/9-1, MU 987/1-1).
1
4
2
2
0.2
0.4
0.6
0.8
1
1.2
1.4
1.6
1.8
x 10
1
0
2
4
2
?2
0.2
0.4
0.6
0.8
1
1.2
1.4
1.6
1.8
x 10
1
?4
2
4
2
?6
?6
?4
?2
0
2
4
0.2
0.4
0.6
0.8
1
1.2
1.4
1.6
1.8
2
4
x 10
Figure 4: A highly nonlinear mixture of two speech signals: Scatterplot of x1 vs x2 and the waveforms of the true source signals (upper panel) in comparison to the
best matching linear and nonlinear separation results are shown in the middle and lower panel, respectively.
References
[1] A. Belouchrani, K. Abed Meraim, J.-F. Cardoso, and E. Moulines. A blind source separation
technique based on second order statistics. IEEE Trans. on Signal Processing, 45(2):434?444,
1997.
[2] G. Burel. Blind separation of sources: a nonlinear neural algorithm.
5(6):937?947, 1992.
Neural Networks,
[3] C.J.C. Burges. A tutorial on support vector machines for pattern recognition. Knowledge
Discovery and Data Mining, 2(2):121?167, 1998.
[4] J.-F. Cardoso. Blind signal separation: statistical principles.
9(10):2009?2025, 1998.
Proceedings of the IEEE,
[5] J.-F. Cardoso and A. Souloumiac. Jacobi angles for simultaneous diagonalization. SIAM
J.Mat.Anal.Appl., 17(1):161 ff., 1996.
[6] N. Cristianini and J. Shawe-Taylor. An Introduction to Support Vector Machines. Cambridge
University Press, Cambridge, UK, 2000.
[7] C. Fyfe and P. L. Lai. ICA using kernel canonical correlation analysis. In Proc. Int. Workshop
on Independent Component Analysis and Blind Signal Separation (ICA2000), pages 279?284,
Helsinki, Finland, 2000.
[8] A. Hyvarinen, J. Karhunen, and E. Oja. Independent Component Analysis. Wiley, 2001.
[9] T.-W. Lee, B.U. Koehler, and R. Orglmeister. Blind source separation of nonlinear mixing
models. In Neural Networks for Signal Processing VII, pages 406?415. IEEE Press, 1997.
[10] J. K. Lin, D. G. Grier, and J. D. Cowan. Faithful representation of separable distributions.
Neural Computation, 9(6):1305?1320, 1997.
[11] G. Marques and L. Almeida. Separation of nonlinear mixtures using pattern repulsion. In Proc.
Int. Workshop on Independent Component Analysis and Signal Separation (ICA?99), pages 277?
282, Aussois, France, 1999.
[12] K.-R. M?ller, S. Mika, G. R?tsch, K. Tsuda, and B. Sch?lkopf. An introduction to kernel-based
learning algorithms. IEEE Transactions on Neural Networks, 12(2):181?201, 2001.
[13] P. Pajunen, A. Hyv?rinen, and J. Karhunen. Nonlinear blind source separation by selforganizing maps. In Proc. Int. Conf. on Neural Information Processing, pages 1207?1210,
Hong Kong, 1996.
[14] P. Pajunen and J. Karhunen. A maximum likelihood approach to nonlinear blind source separation. In Proceedings of the 1997 Int. Conf. on Artificial Neural Networks (ICANN?97), pages
541?546, Lausanne, Switzerland, 1997.
[15] B. Sch?lkopf, S. Mika, C.J.C. Burges, P. Knirsch, K.-R. M?ller, G. R?tsch, and A.J. Smola.
Input space vs. feature space in kernel-based methods. IEEE Transactions on Neural Networks,
10(5):1000?1017, September 1999.
[16] B. Sch?lkopf, A.J. Smola, and K.-R. M?ller. Nonlinear component analysis as a kernel eigenvalue problem. Neural Computation, 10:1299?1319, 1998.
[17] A. Taleb and C. Jutten. Source separation in post-nonlinear mixtures. IEEE Trans. on Signal
Processing, 47(10):2807?2820, 1999.
[18] H. Valpola, X. Giannakopoulos, A. Honkela, and J. Karhunen. Nonlinear independent component analysis using ensemble learning: Experiments and discussion. In Proc. Int. Workshop
on Independent Component Analysis and Blind Signal Separation (ICA2000), pages 351?356,
Helsinki, Finland, 2000.
[19] V.N. Vapnik. The nature of statistical learning theory. Springer Verlag, New York, 1995.
[20] H. H. Yang, S.-I. Amari, and A. Cichocki. Information-theoretic approach to blind separation
of sources in non-linear mixture. Signal Processing, 64(3):291?300, 1998.
[21] A. Ziehe and K.-R. M?ller. TDSEP?an efficient algorithm for blind separation using time
structure. In Proc. Int. Conf. on Artificial Neural Networks (ICANN?98), pages 675?680,
Sk?vde, Sweden, 1998.
| 2094 |@word kong:1 middle:2 polynomial:1 grier:1 underline:1 glue:1 hyv:1 tried:1 pick:1 outperforms:1 existing:1 ida:1 scatter:1 written:1 enables:2 plot:2 v:3 short:1 provides:1 simpler:2 mathematical:2 become:2 ica:3 roughly:1 moulines:1 unfolded:1 str:1 ller1:1 provided:1 moreover:4 bounded:1 panel:9 project:1 what:1 temporal:2 tackle:1 uk:1 before:1 local:1 solely:1 approximately:1 might:1 mika:3 therein:1 challenging:1 appl:1 lausanne:1 faithful:1 practice:2 unfold:1 strasse:1 area:1 projection:1 matching:2 get:3 unfeasible:1 storage:1 ast:1 applying:2 equivalent:1 map:5 formulate:2 array:2 orthonormal:9 spanned:2 retrieve:1 stability:1 handle:1 analogous:1 construction:5 pt:1 rinen:1 trick:3 recognition:1 parameterize:2 eu:1 valuable:2 intuition:1 benjamin:1 mu:1 complexity:1 tsch:3 cristianini:1 efficiency:1 basis:15 compactly:1 easily:1 v22:1 gtm:1 v41:1 separated:2 artificial:2 fyfe:1 klaus:2 choosing:1 quite:1 whose:1 valued:3 amari:1 statistic:1 obviously:1 eigenvalue:1 propose:3 product:5 organizing:1 mixing:5 achieve:1 forth:1 recipe:1 requirement:1 tdsep:14 unmixing:1 derive:1 depending:1 ij:2 eq:9 ddimensional:1 resemble:1 implies:3 indicate:1 motoaki:1 concentrate:1 direction:4 waveform:1 closely:1 switzerland:1 vc:1 centered:1 require:1 ja:1 generalization:1 mathematically:1 extension:1 correction:1 hold:5 exp:5 mapping:2 algorithmic:1 stabilizes:1 finland:2 proc:5 tanh:1 largest:1 agrees:1 successfully:1 establishes:1 stefan:1 clearly:3 sensor:1 gaussian:1 fulfill:2 focus:1 rank:6 likelihood:1 industrial:1 bebel:1 repulsion:1 typically:2 kernelized:2 relation:1 manipulating:1 transformed:1 fhg:1 france:1 germany:2 denoted:1 special:2 field:3 construct:2 having:1 nicely:1 sampling:4 identical:1 unsupervised:1 randomly:1 oja:1 cheaper:1 dfg:1 highly:2 mining:1 mixture:7 sweden:1 machinery:1 taylor:1 tsuda:1 theoretical:1 instance:1 column:13 kekul:1 cost:3 entry:2 expects:1 submanifold:4 emg:2 st:2 siam:1 lee:1 invertible:4 again:1 choose:1 possibly:2 conf:3 strive:1 knirsch:1 toy:3 unmix:1 nonlinearities:2 de:1 sinusoidal:1 coefficient:2 bliss:1 int:6 blind:14 depends:1 vi:4 later:1 doing:1 start:2 pajunen:2 contribution:2 characteristic:1 efficiently:1 ensemble:2 correspond:1 yield:2 lkopf:3 famous:1 meraim:1 simultaneous:2 sebastian:1 definition:1 centering:2 involved:1 naturally:1 hereby:1 jacobi:1 knowledge:2 pulp:1 dimensionality:3 back:1 dt:1 supervised:1 follow:1 specify:1 maximally:1 formulation:2 done:2 though:1 furthermore:2 biomedical:2 smola:2 correlation:6 honkela:1 nonlinear:33 widespread:1 jutten:1 true:3 orthonormality:1 symmetric:1 self:1 coincides:1 hong:1 impression:1 theoretic:1 demonstrate:3 performs:1 reasoning:1 image:1 common:1 khz:1 cambridge:2 similarly:1 nonlinearity:3 shawe:1 dot:4 stable:2 verlag:1 success:1 exploited:1 seen:1 minimum:1 fortunately:1 preserving:1 employed:1 ller:4 signal:30 full:3 match:1 adapt:2 long:1 lin:1 lai:1 post:3 variant:1 kernel:31 achieved:1 addition:1 remarkably:1 addressed:1 source:21 sch:3 elegant:4 thing:1 cowan:1 yang:1 xj:2 identified:1 andreas:1 idea:1 regarding:2 expression:1 pca:2 isomorphism:2 speech:4 york:1 constitute:1 useful:1 cardoso:3 selforganizing:1 repeating:1 svms:1 reduced:3 generate:1 exist:1 canonical:1 tutorial:1 write:1 mat:1 aussois:1 key:1 gunnar:1 d3:1 v1:9 run:2 inverse:1 angle:2 telecommunication:2 vn:1 separation:21 coherence:1 ziehe1:1 correspondence:1 occur:2 x2:5 helsinki:2 u1:2 span:8 extremely:1 formulating:1 performing:1 separable:1 department:1 combination:1 smaller:3 beneficial:1 s31:1 abed:1 s1:5 intuitively:1 ktdsep:12 taken:1 eventually:1 permit:1 apply:2 kawanabe:1 observe:1 kernelizing:1 alternative:1 clustering:2 cf:8 prof:1 establish:1 koehler:1 usual:1 september:1 conceivable:1 subspace:5 thank:1 mapped:7 berlin:1 valpola:1 vd:7 parametrized:1 whom:1 manifold:2 unstable:1 meg:2 length:1 acquire:1 robert:1 anal:1 perform:2 upper:5 finite:1 marque:1 ica2000:2 arbitrary:3 august:1 nonlinearly:3 componentwise:1 acoustic:2 potsdam:2 established:1 trans:2 usually:1 pattern:3 orglmeister:1 power:1 diagonalisation:1 blankertz:1 picture:1 fraunhofer:1 cichocki:1 acknowledgement:1 discovery:1 multiplication:1 sawtooth:1 mixed:5 interesting:1 ingredient:2 remarkable:1 principle:1 viewpoint:1 prone:1 compatible:1 course:1 belouchrani:1 supported:1 last:1 soon:1 allow:1 burges:2 dimension:7 bs:15 calculated:2 world:1 souloumiac:1 giannakopoulos:1 author:1 invertable:1 far:3 hyvarinen:1 transaction:2 confirm:1 assumed:1 xi:2 alternatively:1 sk:1 table:2 nature:1 eeg:2 excellent:1 vj:2 icann:2 main:1 decrement:1 s2:5 big:1 x1:6 fig:4 ff:1 wiley:1 space1:1 lie:3 rk:6 xt:5 demixing:3 intrinsic:3 essential:2 exists:1 scatterplot:3 workshop:3 effectively:2 vapnik:1 diagonalization:1 karhunen:4 vii:1 depicted:1 cheaply:1 visual:1 failed:2 scalar:2 u2:2 springer:1 corresponds:4 relies:2 s32:1 marked:1 consequently:1 rbf:1 manufacturing:1 infinite:2 operates:1 called:1 isomorphic:1 partly:1 ziehe:2 support:4 almeida:1 latter:1 evaluate:1 |
1,201 | 2,095 | Prodding the ROC Curve: Constrained
Optimization of Classifier Performance
Michael C. Mozer*+, Robert Dodier*, Michael D. Colagrosso*+,
C?sar Guerra-Salcedo*, Richard Wolniewicz*
* Advanced Technology Group + Department of Computer Science
Athene Software
University of Colorado
2060 Broadway
Campus Box 430
Boulder, CO 80302
Boulder, CO 80309
Abstract
When designing a two-alternative classifier, one ordinarily aims to maximize the
classifier?s ability to discriminate between members of the two classes. We
describe a situation in a real-world business application of machine-learning
prediction in which an additional constraint is placed on the nature of the solution: that the classifier achieve a specified correct acceptance or correct rejection
rate (i.e., that it achieve a fixed accuracy on members of one class or the other).
Our domain is predicting churn in the telecommunications industry. Churn
refers to customers who switch from one service provider to another. We propose four algorithms for training a classifier subject to this domain constraint,
and present results showing that each algorithm yields a reliable improvement in
performance. Although the improvement is modest in magnitude, it is nonetheless impressive given the difficulty of the problem and the financial return that it
achieves to the service provider.
When designing a classifier, one must specify an objective measure by which the classifier?s performance is to be evaluated. One simple objective measure is to minimize the
number of misclassifications. If the cost of a classification error depends on the target and/
or response class, one might utilize a risk-minimization framework to reduce the expected
loss. A more general approach is to maximize the classifier?s ability to discriminate one
class from another class (e.g., Chang & Lippmann, 1994).
An ROC curve (Green & Swets, 1966) can be used to visualize the discriminative
performance of a two-alternative classifier that outputs class posteriors. To explain the
ROC curve, a classifier can be thought of as making a positive/negative judgement as to
whether an input is a member of some class. Two different accuracy measures can be
obtained from the classifier: the accuracy of correctly identifying an input as a member of
the class (a correct acceptance or CA), and the accuracy of correctly identifying an input
as a nonmember of the class (a correct rejection or CR). To evaluate the CA and CR rates,
it is necessary to pick a threshold above which the classifier?s probability estimate is interpreted as an ?accept,? and below which is interpreted as a ?reject??call this the criterion.
The ROC curve plots CA against CR rates for various criteria (Figure 1a). Note that as the
threshold is lowered, the CA rate increases and the CR rate decreases. For a criterion of 1,
the CA rate approaches 0 and the CR rate 1; for a criterion of 0, the CA rate approaches 1
0
0
correct rejection rate
20
40
60
80
100
100
(b)
correct rejection rate
20
40
60
80
(a)
0
20
40
60
80
100
correct acceptance rate
0
20
40
60
80
100
correct acceptance rate
FIGURE 1. (a) two ROC curves reflecting discrimination performance; the dashed curve
indicates better performance. (b) two plausible ROC curves, neither of which is clearly
superior to the other.
and the CR rate 0. Thus, the ROC curve is anchored at (0,1) and (1,0), and is monotonically nonincreasing. The degree to which the curve is bowed reflects the discriminative
ability of the classifier. The dashed curve in Figure 1a is therefore a better classifier than
the solid curve.
The degree to which the curve is bowed can be quantified by various measures such
as the area under the ROC curve or d?, the distance between the positive and negative distributions. However, training a classifier to maximize either the ROC area or d? often
yields the same result as training a classifier to estimate posterior class probabilities, or
equivalently, to minimize the mean squared error (e.g., Frederick & Floyd, 1998). The
ROC area and d? scores are useful, however, because they reflect a classifier?s intrinsic
ability to discriminate between two classes, regardless of how the decision criterion is set.
That is, each point on an ROC curve indicates one possible CA/CR trade off the classifier
can achieve, and that trade off is determined by the criterion. But changing the criterion
does not change the classifier?s intrinsic ability to discriminate.
Generally, one seeks to optimize the discrimination performance of a classifier. However, we are working in a domain where overall discrimination performance is not as critical as performance at a particular point on the ROC curve, and we are not interested in the
remainder of the ROC curve. To gain an intuition as to why this goal should be feasible,
consider Figure 1b. Both the solid and dashed curves are valid ROC curves, because they
satisfy the monotonicity constraint: as the criterion is lowered, the CA rate does not
decrease and the CR rate does not increase. Although the bow shape of the solid curve is
typical, it is not mandatory; the precise shape of the curve depends on the nature of the
classifier and the nature of the domain. Thus, it is conceivable that a classifier could produce a curve like the dashed one. The dashed curve indicates better performance when the
CA rate is around 50%, but worse performance when the CA rate is much lower or higher
than 50%. Consequently, if our goal is to maximize the CR rate subject to the constraint
that the CA rate is around 50%, or to maximize the CA rate subject to the constraint that
the CR rate is around 90%, the dashed curve is superior to the solid curve. One can imagine that better performance can be obtained along some stretches of the curve by sacrificing performance along other stretches of the curve. Note that obtaining a result such as the
dashed curve requires a nonstandard training algorithm, as the discrimination performance
as measured by the ROC area is worse for the dashed curve than for the solid curve.
In this paper, we propose and evaluate four algorithms for optimizing performance in
a certain region of the ROC curve. To begin, we explain the domain we are concerned with
and why focusing on a certain region of the ROC curve is important in this domain.
1 OUR DOMAIN
Athene Software focuses on predicting and managing subscriber churn in the telecommunications industry (Mozer, Wolniewicz, Grimes, Johnson, & Kaushansky, 2000). ?Churn?
refers to the loss of subscribers who switch from one company to the other. Churn is a significant problem for wireless, long distance, and internet service providers. For example,
in the wireless industry, domestic monthly churn rates are 2?3% of the customer base.
Consequently, service providers are highly motivated to identify subscribers who are dissatisfied with their service and offer them incentives to prevent churn.
We use techniques from statistical machine learning?primarily neural networks and
ensemble methods?to estimate the probability that an individual subscriber will churn in
the near future. The prediction of churn is based on various sources of information about a
subscriber, including: call detail records (date, time, duration, and location of each call,
and whether call was dropped due to lack of coverage or available bandwidth), financial
information appearing on a subscriber?s bill (monthly base fee, additional charges for
roaming and usage beyond monthly prepaid limit), complaints to the customer service
department and their resolution, information from the initial application for service (contract details, rate plan, handset type, credit report), market information (e.g., rate plans
offered by the service provider and its competitors), and demographic data.
Churn prediction is an extremely difficult problem for several reasons. First, the business environment is highly nonstationary; models trained on data from a certain time
period perform far better with hold-out examples from that same time period than examples drawn from successive time periods. Second, features available for prediction are
only weakly related to churn; when computing mutual information between individual
features and churn, the greatest value we typically encounter is .01 bits. Third, information
critical to predicting subscriber behavior, such as quality of service, is often unavailable.
Obtaining accurate churn predictions is only part of the challenge of subscriber
retention. Subscribers who are likely to churn must be contacted by a call center and
offered some incentive to remain with the service provider. In a mathematically principled
business scenario, one would frame the challenge as maximizing profitability to a service
provider, and making the decision about whether to contact a subscriber and what incentive to offer would be based on the expected utility of offering versus not offering an
incentive. However, business practices complicate the scenario and place some unique
constraints on predictive models. First, call centers are operated by a staff of customer service representatives who can contact subscribers at a fixed rate; consequently, our models
cannot advise contacting 50,000 subscribers one week, and 50 the next. Second, internal
business strategies at the service providers constrain the minimum acceptable CA or CR
rates (above and beyond the goal of maximizing profitability). Third, contracts that Athene
makes with service providers will occasionally call for achieving a specific target CA and
CR rate. These three practical issues pose formal problems which, to the best of our
knowledge, have not been addressed by the machine learning community.
The formal problems can be stated in various ways, including: (1) maximize the CA
rate, subject to the constraint that a fixed percentage of the subscriber base is identified as
potential churners, (2) optimize the CR rate, subject to the constraint that the CA rate
should be ?CA, (3) optimize the CA rate, subject to the constraint that the CR rate should
be ?CR, and finally?what marketing executives really want?(4) design a classifier that
has a CA rate of ?CA and a CR rate of ?CR. Problem (1) sounds somewhat different than
problems (2) or (3), but it can be expressed in terms of a lift curve, which plots the CA rate
as a function of the total fraction of subscribers identified by the model. Problem (1) thus
imposes the constraint that the solution lies at one coordinate of the lift curve, just as problems (2) and (3) place the constraint that the solution lies at one coordinate of the ROC
curve. Thus, a solution to problems (2) or (3) will also serve as a solution to (1). Although
addressing problem (4) seems most fanciful, it encompasses problems (2) and (3), and
thus we focus on it. Our goal is not altogether unreasonable, because a solution to problem
(4) has the property we characterized in Figure 1b: the ROC curve can suffer everywhere
except in the region near CA ?CA and CR ?CR. Hence, the approaches we consider will
trade off performance in some regions of the ROC curve against performance in other
regions. We call this prodding the ROC curve.
2 FOUR ALGORITHMS TO PROD THE ROC CURVE
In this section, we describe four algorithms for prodding the ROC curve toward a target
CA rate of ?CA and a target CR rate of ?CR.
2.1 EMPHASIZING CRITICAL TRAINING EXAMPLES
Suppose we train a classifier on a set of positive and negative examples from a class?
churners and nonchurners in our domain. Following training, the classifier will assign a
posterior probability of class membership to each example. The examples can be sorted by
the posterior and arranged on a continuum anchored by probabilities 0 and 1 (Figure 2).
We can identify the thresholds, ?CA and ?CR, which yield CA and CR rates of ?CA and
?CR, respectively. If the classifier?s discrimination performance fails to achieve the target
CA and CR rates, then ?CA will be lower than ?CR, as depicted in the Figure. If we can
bring these two thresholds together, we will achieve the target CA and CR rates. Thus, the
first algorithm we propose involves training a series of classifiers, attempting to make classifier n+1 achieve better CA and CR rates by focusing its effort on examples from classifier n that lie between ?CA and ?CR; the positive examples must be pushed above ?CR and
the negative examples must be pushed below ?CA. (Of course, the thresholds are specific
to a classifier, and hence should be indexed by n.) We call this the emphasis algorithm,
because it involves placing greater weight on the examples that lie between the two thresholds. In the Figure, the emphasis for classifier n+1 would be on examples e5 through e8.
This retraining procedure can be iterated until the classifier?s training set performance
reaches asymptote.
In our implementation, we define a weighting of each example i for training classifier
n, ? in . For classifier 1, ? i1 = 1 . For subsequent classifiers, ? in + 1 = ? in if example i is
not in the region of emphasis, or ? in + 1 = ? e ? in otherwise, where ?e is a constant, ?e > 1.
2.2 DEEMPHASIZING IRRELEVANT TRAINING EXAMPLES
The second algorithm we propose is related to the first, but takes a slightly different perspective on the continuum depicted in Figure 2. Positive examples below ?CA?such as
e2?are clearly the most dif ficult positive examples to classify correctly. Not only are they
the most difficult positive examples, but they do not in fact need to be classified correctly
to achieve the target CA and CR rates. Threshold ?CR does not depend on examples such
as e2, and threshold ?CA allows a fraction (1??CA) of the positive examples to be classified
incorrectly. Likewise, one can argue that negative examples above ?CR?such as e10 and
e11?need not be of concern. Essentially , the second algorithm, which we term thedeemphasis algorithm, is like the emphasis algorithm in that a series of classifiers are trained,
but when training classifier n+1, less weight is placed on the examples whose correct clas?CA
e1 e2 e3
0
e4
?CR
e5
e6
e7
e8
churn probability
e9
e10 e11 e12 e13
1
FIGURE 2. A schematic depiction of all training examples arranged by the classifier?s
posterior. Each solid bar corresponds to a positive example (e.g., a churner) and each grey bar
corresponds to a negative example (e.g., a nonchurner).
sification is unnecessary to achieve the target CA and CR rates for classifier n. As with the
emphasis algorithm, the retraining procedure can be iterated until no further performance
improvements are obtained on the training set. Note that the set of examples given emphasis by the previous algorithm is not the complement of the set of examples deemphasized
by the current algorithm; the algorithms are not identical.
In our implementation, we assign a weight to each example i for training classifier n,
? in . For classifier 1, ? i1 = 1 . For subsequent classifiers, ? in + 1 = ? in if example i is not
in the region of deemphasis, or ? in + 1 = ? d ? in otherwise, where ?d is a constant, ?d <1.
2.3 CONSTRAINED OPTIMIZATION
The third algorithm we propose is formulated as maximizing the CR rate while maintaining the CA rate equal to ?CA. (We do not attempt to simultaneously maximize the CA rate
while maintaining the CR rate equal to ?CR.) Gradient methods cannot be applied directly
because the CA and CR rates are nondifferentiable, but we can approximate the CA and
CR rates with smooth differentiable functions:
1
1
CA ( w, t ) = ------ ? ? ? ( f ( x i, w ) ? t )
CR ( w, t ) = ------- ? ? ? ( t ? f ( x i, w ) ) ,
P i?P
N i?N
where P and N are the set of positive and negative examples, respectively, f(x,w) is the
model posterior for input x, w is the parameterization of the model, t is a threshold, and ??
?1
is a sigmoid function with scaling parameter ?: ? ? ( y ) = ( 1 + exp ( ? ?y ) ) . The larger ?
is, the more nearly step-like the sigmoid is and the more nearly equal the approximations
are to the model CR and CA rates. We consider the problem formulation in which CA is a
constraint and CR is a figure of merit. We convert the constrained optimization problem
into an unconstrained problem by the augmented Lagrangian method (Bertsekas, 1982),
which involves iteratively maximizing an objective function
2
?
A ( w, t ) = CR ( w, t ) + ? CA ( w, t ) ? ? CA + --- CA ( w, t ) ? ? CA
2
with a fixed Lagrangian multiplier, ?, and then updating ? following the optimization step:
? ? ? + ? CA ( w *, t * ) ? ? CA , where w * and t * are the values found by the optimization
step. We initialize ? = 1 and fix ? = 1 and ? = 10 and iterate until ? converges.
2.4 GENETIC ALGORITHM
The fourth algorithm we explore is a steady-state genetic search over a space defined by
the continuous parameters of a classifier (Whitley, 1989). The fitness of a classifier is the
reciprocal of the number of training examples falling between the ?CA and ?CR thresholds.
Much like the emphasis algorithm, this fitness function encourages the two thresholds to
come together. The genetic search permits direct optimization over a nondifferentiable criterion, and therefore seems sensible for the present task.
3 METHODOLOGY
For our tests, we studied two large data bases made available to Athene by two telecommunications providers. Data set 1 had 50,000 subscribers described by 35 input features
and a churn rate of 4.86%. Data set 2 had 169,727 subscribers described by 51 input features and a churn rate of 6.42%. For each data base, the features input to the classifier were
obtained by proprietary transformations of the raw data (see Mozer et al., 2000). We chose
these two large, real world data sets because achieving gains with these data sets should be
more difficult than with smaller, less noisy data sets. Plus, with our real-world data, we
can evaluate the cost savings achieved by an improvement in prediction accuracy. We performed 10-fold cross-validation on each data set, preserving the overall churn/nonchurn
ratio in each split.
In all tests, we chose ? CR = 0.90 and ? CA = 0.50 , values which, based on our
past experience in this domain, are ambitious yet realizable targets for data sets such as
these. We used a logistic regression model (i.e., a no hidden unit neural network) for our
studies, believing that it would be more difficult to obtain improvements with such a
model than with a more flexible multilayer perceptron. For the emphasis and deemphasis
algorithms, models were trained to minimize mean-squared error on the training set. We
chose ?e = 1.3 and ?d = .75 by quick exploration. Because the weightings are cumulative
over training restarts, the choice of ? is not critical for either algorithm; rather, the magnitude of ? controls how many restarts are necessary to reach asymptotic performance, but
the results we obtained were robust to the choice of ?. The emphasis and deemphasis algorithms were run for 100 iterations, which was the number of iterations required to reach
asymptotic performance on the training set.
4 RESULTS
Figure 3 illustrates training set performance for the emphasis algorithm on data set 1. The
graph on the left shows the CA rate when the CR rate is .9, and the graph on the right show
the CR rate when the CA rate is .5. Clearly, the algorithm appears to be stable, and the
ROC curve is improving in the region around (?CA, ?CR).
Figure 4 shows cross-validation performance on the two data sets for the four prodding algorithms as well as for a traditional least-squares training procedure. The emphasis
and deemphasis algorithms yield reliable improvements in performance in the critical
region of the ROC curve over the traditional training procedure. The constrained-optimization and genetic algorithms perform well on achieving a high CR rate for a fixed CA
rate, but neither does as well on achieving a high CA rate for a fixed CR rate. For the constrained-optimization algorithm, this result is not surprising as it was trained asymmetrically, with the CA rate as the constraint. However, for the genetic algorithm, we have little
explanation for its poor performance, other than the difficulty faced in searching a continuous space without gradient information.
5 DISCUSSION
0.4
0.845
0.395
0.84
0.39
0.835
0.385
CR rate
CA rate
In this paper, we have identified an interesting, novel problem in classifier design which is
motivated by our domain of churn prediction and real-world business considerations.
Rather than seeking a classifier that maximizes discriminability between two classes, as
measured by area under the ROC curve, we are concerned with optimizing performance at
certain points along the ROC curve. We presented four alternative approaches to prodding
the ROC curve, and found that all four have promise, depending on the specific goal.
Although the magnitude of the gain is small?an increase of about .01 in the CR rate
given a target CA rate of .50?the impro vement results in significant dollar savings. Using
a framework for evaluating dollar savings to a service provider, based on estimates of subscriber retention and costs of intervention obtained in real world data collection (Mozer et
0.38
0.83
0.825
0.375
0.82
0.37
0.815
0.365
0.81
0
5
10 15 20 25 30 35 40 45 50
Iteration
0
5
10 15 20 25 30 35 40 45 50
Iteration
FIGURE 3. Training set performance for the emphasis algorithm on data set 1. (a) CA rate as
a function of iteration for a CR rate of .9; (b) CR rate as a function of iteration for a CA rate of
.5. Error bars indicate +/?1 standard error of the mean.
0.840
0.385
0.835
0.380
0.830
0.375
0.825
CR rate
Data set 1
CA rate
ISP
Test Set
0.390
0.370
0.820
0.365
0.815
0.360
0.810
0.355
0.805
0.350
0.800
std
emph
deemph
constr
GA
std
emph
deemph
constr
GA
std
emph
deemph
constr
GA
0.900
0.375
0.350
CR rate
Data set 2
0.875
CA rate
Wireless
Test Set
0.850
0.325
0.825
0.300
0.800
std
emph
deemph
constr
GA
FIGURE 4. Cross-validation performance on the two data sets for the standard training
procedure (STD), as well as the emphasis (EMPH), deemphasis (DEEMPH), constrained
optimization (CONSTR), and genetic (GEN) algorithms. The left column shows the CA rate for
CR rate .9; the right column shows the CR rate for CA rate .5. The error bar indicates one
standard error of the mean over the 10 data splits.
al., 2000), we obtain a savings of $11 per churnable subscriber when the (CA, CR) rates
go from (.50, .80) to (.50, .81), which amounts to an 8% increase in profitability of the
subscriber intervention effort.
These figures are clearly promising. However, based on the data sets we have studied, it is difficult to know whether another algorithm might exist that achieves even greater
gains. Interestingly, all algorithms we proposed yielded roughly the same gains when successful, suggesting that we may have milked the data for whatever gain could be had,
given the model class evaluated. Our work clearly illustrate the difficulty of the problem,
and we hope that others in the NIPS community will be motivated by the problem to suggest even more powerful, theoretically grounded approaches.
6 ACKNOWLEDGEMENTS
No white males were angered in the course of conducting this research. We thank Lian Yan and
David Grimes for comments and assistance on this research. This research was supported in part by
McDonnell-Pew grant 97-18, NSF award IBN-9873492, and NIH/IFOPAL R01 MH61549?01A1.
7 REFERENCES
Bertsekas, D. P. (1982). Constrained optimization and Lagrange multiplier methods. NY: Academic.
Chang, E. I., & Lippmann, R. P. (1994). Figure of merit training for detection and spotting. In J. D.
Cowan, G. Tesauro, & J. Alspector (Eds.), Advances in Neural Information Processing Systems
6 (1019?1026). San Mateo, CA: Morgan Kaufmann.
Frederick, E. D., & Floyd, C. E. (1998). Analysis of mammographic findings and patient history
data with genetic algorithms for the prediction of breast cancer biopsy outcome. Proceedings of
the SPIE, 3338, 241?245.
Green, D. M., & Swets, J. A. (1966). Signal detection theory and psychophysics. New York: Wiley.
Mozer, M. C., Wolniewicz, R., Grimes, D., Johnson, E., & Kaushansky, H. (2000). Maximizing revenue by predicting and addressing customer dissatisfaction. IEEE Transactions on Neural Networks, 11, 690?696.
Whitley, D. (1989). The GENITOR algorithm and selective pressure: Why rank-based allocation of
reproductive trials is best. In D. Schaffer (Ed.), Proceedings of the Third International Conference on Genetic Algorithms (pp. 116?121). San Mateo, CA: Morgan Kaufmann.
| 2095 |@word trial:1 judgement:1 seems:2 retraining:2 grey:1 seek:1 subscriber:19 salcedo:1 pick:1 pressure:1 solid:6 initial:1 series:2 score:1 offering:2 genetic:8 interestingly:1 past:1 current:1 surprising:1 yet:1 must:4 subsequent:2 shape:2 asymptote:1 plot:2 discrimination:5 parameterization:1 reciprocal:1 record:1 location:1 bowed:2 successive:1 along:3 direct:1 contacted:1 theoretically:1 swets:2 market:1 expected:2 roughly:1 alspector:1 behavior:1 company:1 little:1 domestic:1 begin:1 campus:1 maximizes:1 what:2 interpreted:2 finding:1 transformation:1 charge:1 complaint:1 classifier:47 control:1 unit:1 whatever:1 intervention:2 grant:1 bertsekas:2 positive:10 service:15 dropped:1 retention:2 limit:1 might:2 chose:3 emphasis:13 plus:1 studied:2 quantified:1 discriminability:1 mateo:2 co:2 dif:1 unique:1 practical:1 practice:1 procedure:5 area:5 yan:1 thought:1 reject:1 refers:2 advise:1 suggest:1 cannot:2 ga:4 risk:1 optimize:3 bill:1 customer:5 center:2 maximizing:5 lagrangian:2 quick:1 regardless:1 go:1 duration:1 resolution:1 identifying:2 financial:2 searching:1 coordinate:2 sar:1 imagine:1 target:10 colorado:1 suppose:1 designing:2 dodier:1 updating:1 std:5 dissatisfied:1 region:9 decrease:2 trade:3 e8:2 principled:1 mozer:5 intuition:1 environment:1 trained:4 weakly:1 depend:1 predictive:1 serve:1 isp:1 various:4 train:1 describe:2 lift:2 outcome:1 whose:1 larger:1 plausible:1 whitley:2 otherwise:2 ability:5 noisy:1 differentiable:1 propose:5 remainder:1 date:1 bow:1 gen:1 achieve:8 produce:1 converges:1 depending:1 illustrate:1 pose:1 measured:2 coverage:1 ibn:1 involves:3 come:1 indicate:1 biopsy:1 correct:9 exploration:1 assign:2 fix:1 really:1 mathematically:1 stretch:2 hold:1 around:4 credit:1 exp:1 week:1 visualize:1 achieves:2 continuum:2 reflects:1 minimization:1 hope:1 clearly:5 aim:1 e7:1 rather:2 cr:60 focus:2 improvement:6 rank:1 indicates:4 believing:1 realizable:1 dollar:2 mh61549:1 membership:1 typically:1 accept:1 hidden:1 genitor:1 selective:1 interested:1 i1:2 e11:2 overall:2 classification:1 issue:1 flexible:1 plan:2 constrained:7 initialize:1 mutual:1 mammographic:1 equal:3 psychophysics:1 saving:4 identical:1 placing:1 nearly:2 future:1 report:1 others:1 richard:1 primarily:1 handset:1 simultaneously:1 individual:2 fitness:2 attempt:1 detection:2 acceptance:4 highly:2 male:1 grime:3 operated:1 nonincreasing:1 accurate:1 necessary:2 experience:1 modest:1 indexed:1 sacrificing:1 industry:3 classify:1 column:2 cost:3 addressing:2 successful:1 johnson:2 international:1 contract:2 off:3 michael:2 together:2 deemphasizing:1 squared:2 reflect:1 e9:1 worse:2 return:1 colagrosso:1 suggesting:1 potential:1 ifopal:1 churnable:1 satisfy:1 depends:2 performed:1 e13:1 minimize:3 square:1 accuracy:5 kaufmann:2 who:5 likewise:1 ensemble:1 yield:4 identify:2 conducting:1 raw:1 iterated:2 provider:11 churn:19 classified:2 e12:1 nonstandard:1 explain:2 history:1 reach:3 complicate:1 ed:2 against:2 competitor:1 nonetheless:1 pp:1 e2:3 spie:1 gain:6 knowledge:1 reflecting:1 focusing:2 appears:1 higher:1 athene:4 restarts:2 methodology:1 specify:1 response:1 arranged:2 evaluated:2 box:1 formulation:1 profitability:3 marketing:1 just:1 until:3 working:1 lack:1 logistic:1 quality:1 usage:1 multiplier:2 hence:2 iteratively:1 white:1 floyd:2 assistance:1 encourages:1 steady:1 criterion:9 bring:1 consideration:1 novel:1 nih:1 superior:2 sigmoid:2 significant:2 monthly:3 pew:1 unconstrained:1 had:3 lowered:2 stable:1 impressive:1 depiction:1 base:5 posterior:6 perspective:1 optimizing:2 irrelevant:1 tesauro:1 mandatory:1 scenario:2 certain:4 occasionally:1 preserving:1 minimum:1 additional:2 somewhat:1 staff:1 greater:2 morgan:2 managing:1 maximize:7 period:3 monotonically:1 dashed:8 signal:1 sound:1 smooth:1 characterized:1 academic:1 offer:2 long:1 cross:3 e1:1 award:1 a1:1 wolniewicz:3 prediction:8 schematic:1 regression:1 multilayer:1 essentially:1 patient:1 breast:1 iteration:6 grounded:1 achieved:1 want:1 addressed:1 source:1 comment:1 subject:6 cowan:1 member:4 call:9 nonstationary:1 near:2 split:2 concerned:2 switch:2 iterate:1 misclassifications:1 bandwidth:1 identified:3 reduce:1 whether:4 motivated:3 utility:1 effort:2 suffer:1 e3:1 york:1 proprietary:1 useful:1 generally:1 amount:1 percentage:1 exist:1 nsf:1 correctly:4 per:1 promise:1 incentive:4 group:1 four:7 threshold:11 achieving:4 drawn:1 falling:1 changing:1 prevent:1 neither:2 utilize:1 graph:2 nonchurners:1 fraction:2 convert:1 nonchurn:1 run:1 everywhere:1 fourth:1 telecommunication:3 powerful:1 place:2 decision:2 acceptable:1 scaling:1 fee:1 bit:1 pushed:2 internet:1 fold:1 yielded:1 constraint:13 constrain:1 software:2 churners:2 extremely:1 attempting:1 department:2 mcdonnell:1 poor:1 remain:1 slightly:1 smaller:1 constr:5 making:2 boulder:2 know:1 merit:2 demographic:1 available:3 permit:1 unreasonable:1 appearing:1 alternative:3 encounter:1 altogether:1 deemphasized:1 maintaining:2 r01:1 contact:2 seeking:1 objective:3 contacting:1 strategy:1 traditional:2 conceivable:1 gradient:2 distance:2 thank:1 sensible:1 nondifferentiable:2 argue:1 reason:1 toward:1 ratio:1 equivalently:1 difficult:5 robert:1 broadway:1 negative:7 stated:1 ordinarily:1 design:2 implementation:2 ambitious:1 perform:2 incorrectly:1 situation:1 precise:1 frame:1 community:2 schaffer:1 david:1 complement:1 required:1 specified:1 nip:1 frederick:2 beyond:2 bar:4 below:3 spotting:1 guerra:1 challenge:2 encompasses:1 reliable:2 green:2 including:2 explanation:1 greatest:1 critical:5 business:6 difficulty:3 predicting:4 advanced:1 technology:1 faced:1 acknowledgement:1 asymptotic:2 loss:2 interesting:1 allocation:1 versus:1 validation:3 executive:1 revenue:1 degree:2 offered:2 imposes:1 cancer:1 course:2 placed:2 wireless:3 supported:1 prodding:5 formal:2 perceptron:1 sification:1 roaming:1 curve:44 world:5 valid:1 cumulative:1 evaluating:1 made:1 collection:1 san:2 far:1 transaction:1 approximate:1 lippmann:2 monotonicity:1 unnecessary:1 discriminative:2 search:2 continuous:2 prod:1 anchored:2 why:3 promising:1 nature:3 robust:1 ca:73 obtaining:2 unavailable:1 improving:1 e5:2 domain:10 kaushansky:2 augmented:1 representative:1 roc:28 ny:1 wiley:1 fails:1 lie:4 third:4 weighting:2 e4:1 emphasizing:1 specific:3 showing:1 reproductive:1 concern:1 intrinsic:2 magnitude:3 illustrates:1 rejection:4 depicted:2 likely:1 clas:1 explore:1 dissatisfaction:1 lagrange:1 expressed:1 chang:2 ficult:1 corresponds:2 goal:5 sorted:1 formulated:1 consequently:3 feasible:1 change:1 determined:1 typical:1 except:1 total:1 discriminate:4 asymmetrically:1 e10:2 internal:1 e6:1 evaluate:3 lian:1 |
1,202 | 2,096 | Bayesian time series classification
Peter Sykacek
Department of Engineering Science
University of Oxford
Oxford, OX1 3PJ, UK
[email protected]
Stephen Roberts
Department of Engineering Science
University of Oxford
Oxford, OX1 3PJ, UK
[email protected]
Abstract
This paper proposes an approach to classification of adjacent segments
of a time series as being either of classes. We use a hierarchical model
that consists of a feature extraction stage and a generative classifier which
is built on top of these features. Such two stage approaches are often used
in signal and image processing. The novel part of our work is that we link
these stages probabilistically by using a latent feature space. To use one
joint model is a Bayesian requirement, which has the advantage to fuse
information according to its certainty.
The classifier is implemented as hidden Markov model with Gaussian
and Multinomial observation distributions defined on a suitably chosen
representation of autoregressive models. The Markov dependency is motivated by the assumption that successive classifications will be correlated. Inference is done with Markov chain Monte Carlo (MCMC) techniques. We apply the proposed approach to synthetic data and to classification of EEG that was recorded while the subjects performed different
cognitive tasks. All experiments show that using a latent feature space
results in a significant improvement in generalization accuracy. Hence
we expect that this idea generalizes well to other hierarchical models.
1 Introduction
Many applications in signal or image processing are hierarchical in the sense that a probabilistic model is built on top of variables that are the coefficients of some feature extraction
technique. In this paper we consider a particular problem of that kind, where a Gaussian
and Multinomial observation hidden Markov model (GMOHMM) is used to discriminate
coefficients of an Auto Regressive (AR) process as being either of
classes. Bayesian
inference is known to give reasonable results when applied to AR models ([RF95]). The
situation with classification is similar, see for example the seminal work by [Nea96] and
[Mac92]. Hence we may expect to get good results if we apply Bayesian techniques to
both stages of the decision process separately. However this is suboptimal since it meant
to establish a no probabilistic link between feature extraction and classification. Two arguments suggest the building of one probabilistic model which combines feature extraction
and classification:
Since there is a probabilistic link, the generative classifier acts as a prior for fea-
ture extraction. The advantage of using this prior is that it naturally encodes our
knowledge about features as obtained from training data and other sensors. Obviously this is the only setup that is consistent with Bayesian theory ([BS94]).
Since all inferences are obtained from marginal distributions, information is combined according to its certainty. Hence we expect to improve results since information from different sensors is fused in an optimal manner.
2 Methods
2.1 A Gaussian and Multinomial observation hidden Markov model
As we attempt to classify adjacent segments of a time series, it is very likely that we find
correlations between successive class labels. Hence our model has a hidden Markov model
([RJ86]) like architecture, with diagonal Gaussian observation models for continuous variables and Multinomial observation models for discrete variables. We call the architecture a Gaussian and Multinomial observation hidden Markov model or GMOHMM for
short. Contrary to the classical approach, where each class is represented by its own trained
HMM, our model has class labels which are child nodes of the hidden state variables. Figure 1 shows the directed acyclic graph (DAG) of our model. We use here the convention
found in [RG97], where circular nodes are latent and square nodes are observed variables.
2.1.1 Quantities of interest
We regard all variables in the DAG that represent the probabilistic model of the time series
as quantities of interest. These are the hidden states, , the variables of the latent feature
space, , and , the class labels, , the sufficient statistics of the AR process,
,
and the segments of the time series, . The DAG shows the observation model only for
the
-th state. We have
latent feature variables, , which represent the coefficients of
the preprocessing model for of the
-th segment at sensor
. The state conditional distributions, , are modeled by diagonal Gaussians. Variable is the latent model
indicator which represents the model order of the preprocessing model and hence the dimension of . The corresponding observation model is a Multinomial-one
distribution. The third observation, , represents the class label of the
-th segment. The
observation model for is again a Multinomial-one distribution. Note that depending on whether we know the class label or not, can be a latent variable or observed. The
child node of and is the observed variable
, which represents a sufficient statistics of the corresponding time series segment. The proposed approach requires to calculate
the likelihoods !
repeatedly. Hence using the sufficient statistics is a
computational necessity. Finally we use " to represent the precision of the residual noise
model. The noise level is a nuisance parameter which is integrated over.
2.1.2 Model coefficients
Since we integrate over all unknown quantities, there is no conceptual difference between
model coefficients and the variables described above. However there is a qualitative difference. Model parameters exist only once for the entire GMOHMM, whereas there is an
individual quantity of interest for every segment
. Furthermore the model coefficients are
only updated during model inference whereas all quantities of interest are updated during
model inference and for prediction. We have three different prior counts, #%$ , #& and #' ,
which define the Dirichlet priors of the corresponding probabilities. Variable ( denotes
the transition probabilities, that is )* ,+- /. (1032 . The model assumes a stationary hidden state sequence. This allows us to obtain the unconditional prior probability of states 4
from the recurrence relation 56$78 /. (1598:;- . The prior probability of the first hidden
state, 5 $ - , is therefore the normalized eigenvector of the transition probability matrix
( that corresponds to the eigenvalue . Variable
represents the probabilities of class ,
)* .
032 , which are conditional on as well. The prior probabilities for observing the model indicator are represented by 5 . The probability )* 8 . 5 0 2 is
again conditional on the state . As was mentioned above, represents the model order
of the time series model. Hence another interpretation of 5 is that of state dependent prior
probabilities for observing particular model orders. The observation models for are
dynamic mixtures of Gaussians, with one model for each sensor
. Variables and 1
represent the coefficients of all Gaussian kernels. Hence 1 8 is a variate Gaussian distribution. Another interpretation is that the discrete indicator variables
8 and determine together with and 1 a Gaussian prior over . The nodes ,
, , , and / define a hierarchical prior setting which is discussed below.
?
T
T
d
t
d
i?1
?
W
i
d
i
W
i+1
?
?
1
?
?
s
P
?
1
?1
?
?1
i,1
I
i,1
P
1
?
1
h
?
?
i,s
I
i,s
P
s
?
?
i,s
?s
?s
1
X
s
s
i,1
g
P
s
?i,1
?
s
?
?
?s
?1
?1
g
s
h
s
X
i,1
i,s
i,s
Figure 1: This figure illustrates the details of the proposed model as a directed acyclic
graph. The graph shows the model parameters and all quantities of interest: denotes the
hidden states of the HMM; are the class labels of the corresponding time series segments;
are the latent coefficients of the time series model and the corresponding model
indicator variables; is the precision of the residual noise. For tractable inference, we
extract from the time series the sufficient statistics
. All other variables denote
model coefficients: ( are the transition probabilities;
are the probabilities for class 3 ;
and 1 are mean vectors and covariance matrices of the Gaussian observation model
for sensor
; and ) are the probabilities for observing .
2.2 Likelihood and priors for the GMOHMM
Suppose that we are provided with
segments of training data, .
. The
likelihood function of the GMOHMM parameters is then obtained by summation over all
possible sequences, , of latent states, . The sums and integrals under the product make
the likelihood function of Equation (1) highly nonlinear. This may be resolved by using
Gibbs sampling [GG84], which uses tricks similar to those of the expectation maximization
algorithm.
)*
/. )* - -
-
)* : -
- 2
"!
#$
- -3 "- -3 - -3 "-
7
(1)
%&'
(
Gibbs sampling requires that we obtain full conditional distributions1 we can sample from.
The conjugate priors are adopted from [RG97]. Below square brackets and index are
used to denote a particular component of a vector or matrix. Each component mean, 032 ,
- : - , with denoting the mean and : - the
is given a Gaussian prior: 0 2
inverse covariance matrix. As we use diagonal covariance matrices, we may give each
diagonal element an independent Gamma prior: * 0 2 :; , where
denotes the shape parameter and
denotes the inverse scale parameter. The hyperparameter, ! , gets a component wise Gamma hyper prior:
/ . The state
032 , get a Dirichlet prior:
032
# & # & . The
conditional class probabilities,
transition probabilities, ( 0 2 , get a Dirichlet prior: ( 0 2
#$ #$ . The probabilities
for observing different model orders, 5 3 0 2 , depend on the state . Their prior is Dirichlet
5 032
# ' # '/ . The precision gets a Jeffreys? prior, i.e. the scale parameter
is set to 0.
2
*),+
- (0/
.- ( (0/ )1 - (0/
- (0/ )31
4- (0' /'
' ' )65
)75
:<;
)85 '9'
and = , B 8 is set between > ' = and and
/?- (0/ is typically beValues Afor
are between
A
@
B
A
@
B
!
tween
- (?/ and > - (?/ ! , with - (0/ denoting the input range of maximum likelihood
estimates for - (0/
. The mean, , is the midpoint of the maximum likelihood esti4@?B C- (0/ ! , where B A- (0/ is again
mates - (0/
. The inverse covariance matrix A- (?/ .
the range of the estimates at sensor
. We set the prior counts #&
and #$ and #' to .
2.3 Sampling from the posterior
During model inference we need to update all unobserved variables of the DAG, whereas
for predictions we update only the variables summarized in section 2.1.1. Most of the
updates are done using the corresponding full conditional distributions, which have the
same functional forms as the corresponding priors. These full conditionals follow closely
from what was published previously in [Syk00], with some modifications necessary (see
e.g. [Rob96]), because we need to consider the Markov dependency between successive
hidden states. As the derivations of the full conditionals do not differ much from previous
work, we will omit them here and instead concentrate on an illustration how to update the
latent feature space,
.
2.3.1 A representation of the latent feature space
:<D
The AR model in Equation (2) is a linear regression model. We use 2 to denote the
AR coefficients, to denote the model order and to denote a sample from the noise
process, which we assume to be Gaussian with precision .
E4- /
F - / .HG 2 : D 2 F - GJI L/ KE4- /
(2)
D -
As is indicated by the subscript , the value of the I -th AR coefficient depends on the
model order. Hence AR coefficients are not a convenient representation of the latent feature
1
These are the distributions obtained when we condition on all other variables of the DAG.
space. A much more convenient representation is provided by using reflection coefficients,
, (statistically speaking they are partial correlation coefficients), which relate to AR coefficients via
the order recursive Levinson algorithm. Below we use vector notation and the
symbol 2 to denote the upside down version of the AR coefficient vector.
2 +-
K
2
(3)
2 +- .
2 +-
from dynamically
We expect to observe only such data that was generated
pro 2 stable AR
2 . This
cesses. For such processes, the latent density is defined on
- G /
as probais in contrast with the proposed DAG, where we use a finite Gaussian mixture
2
bilistic model for the latent variable, which is is defined on
. In order to avoid
applying
! ,
this mismatch, we reparameterise the space of reflection coefficients by
2
to obtain a more convenient representation of the latent features.
. !;"
(4)
2.3.2 Within dimensional updates
The within dimensional updates can be done with a conventional Metropolis Hastings step.
Integrating out , we obtain a Student t distributed likelihood function of the AR coefficients. In order to obtain likelihood ratio 1, we propose from the multivariate Student-t
distribution shown below, reparameterise in terms of reflection coefficients and apply the
! transformation.
$#
.
!;"
/%&#
where
)
' (*)
.
+
.
+
#
with
)
-
:;-
G
.
- B ' '9' B
:;-,
B *G
(5)
, $ :;-,
+
-
=
B
The proposal uses + to denote the -dimensional sample auto-covariance matrix,
is
,
the sample variance, .
- 2 +- $ is a vector of sample autocorrelations at lags
to
and N denotes the number of samples of the time series . The proposal in
Equation (5) gives a likelihood ratio of . The corresponding acceptance probability is
K
/
92
# 6
5 92
5
587 2
5 7 2
5
:<;
5
(6)
555 = '
5 7 5
The determinant of the Jacobian arises because we7 transform the AR coefficients using
:
..0/ 2134
5
Equations (3) and (4).
2.3.3 Updating model orders
Updating model orders requires us to sample across different dimensional parameter
spaces. One way of doing this is by using the reversible jump MCMC which was recently
proposed in [Gre95]. > We implement the reversible jump move from parameter space
>
3 2 to parameter space 2 +- as partial proposal. That is we propose a reflection
coefficient from a distribution that is conditional on the AR coefficient / . Integrating
out the precision of the noise model ! we obtain again a Student-t distributed likelihood.
This suggests the following proposal:
#
.
where
-
)
' (
(7)
G !-
.
.
G !
= G
B G $ , $
B K = , $ K2 + $
-
/
with
!;
.
.
"! K = K2 + * '
statistics of the K -dimensional BAR process,
Equation B (7) makes
use of the sufficient
B
. ' ' +"! . We use to denote the number
and
to denote
B +- / and
, $ ofB observations
the estimated auto covariance at time lag to obtain . - - ' '
+ as dimensional sample covariance matrix. Assuming
that> the probability of
proposing this move
>
is independent of , the proposal from 3 2 to 2 +- has acceptance probability
:;-
:
!
: . . / G !-!
= 1 1 !! # G ! K 8' (8)
>
>
If we attempt an update from 2 +- to 2 , we have to invert the second argument
of the .0/ operation in Equation (8).
!
.
+
3 Experiments
Convergence of all experiments is analysed by applying the method suggested in [RL96]
to the sequence of observed data likelihoods (equation (1), when filling in all variables).
3.1 Synthetic data
Our first evaluation uses synthetic data. We generate a first order Markov sequence as target
labels (2 state values) with 200 samples used for training and 600 used for testing. Each
sample is used as label of a segment with 200 samples from an auto regressive process. If
. If the label is
the label is , we generate data using reflection coefficients
. The driving noise has variance . Due to sampling
, we use the model
effects we obtain a data set with Bayes error
. In order to make the problem more
realistic, we use a second state sequence to replace of the segments with white noise.
These ?artifacts? are not correlated with the class labels.
=
>' G >' >'
>
> ' G > ' > '
=0>
In order to assess the effect of using a latent feature space, we perform three different
tests: In the first run we use conventional feature extraction with a third order model and
estimates found with maximum likelihood; In a second run we use again a third order
model but integrate over feature values; Finally the third test uses the proposed architecture
with a prior over model order which is ?flat? between and .
>
When compared with conditioning on feature estimates, the latent features show increased
likelihood. The likelihood gets even larger when we regard both the feature values and the
model orders of the preprocessing stage as random variables. As can be seen in figure 2,
this effect is also evident when we look at the generalization probabilities which become
larger as well. We explain this by sharper ?priors? over feature values and model orders,
which are due to the information provided by temporal context 2 of every segment. This
reduces the variance of the observation models which in turn increases likelihoods and
target probabilities. Table 1 shows that these higher probabilities correspond to a significant
improvement in generalization accuracy.
Probabilities from conditioning
1
0.5
0
50
100
50
100
50
100
150
200
250
300
350
400
Probabilities from integrating over features
450
500
550
600
150
200
250
300
350
400
450
500
Probabilities from integrating over model orders and features
550
600
550
600
1
0.5
0
1
0.5
0
150
200
250
300
350
400
450
500
Figure 2: This figure shows the generalization probabilities obtained with different settings. We see that the class probabilities get larger when we regard features as random
variables. This effect is even stronger when both the features and the model orders are
random variables.
3.2 Classification of cognitive tasks
The data used in these experiments is EEG recorded from 5 young, healthy and untrained
subjects while they perform different cognitive tasks. We classify 2 task pairings: auditorynavigation and left motor-right motor imagination. The recordings were taken from 3 electrode sites: T4, P4 (right tempero-parietal for spatial and auditory tasks), C3? , C3? (left
motor area for right motor imagination) and C4? , C4? (right motor area for left motor
imagination). The ground electrode was placed just lateral to the left mastoid process. The
data were recorded using an ISO-DAM system (gain of and fourth order band pass
filter with pass band between Hz and Hz). These signals were sampled with 384
Hz and 12 bit resolution. Each cognitive experiment was performed times for
seconds.
Classification uses again the same settings as with the synthetic problem. The summary
in table 1 shows results obtained from fold cross validation, where one experiment is
used for testing whereas all remaining data is used for training. We observe again significantly improved results when we regard features and model orders as latent variables.
The values in brackets are the significance levels for comparing integration of features with
conditioning and full integration with integration over feature values only.
4 Discussion
We propose in this paper a novel approach to hierarchical time series processing which
makes use of a latent feature representation. This understanding of features and model
orders as random variables is a direct consequence of applying Bayesian theory. Empirical
2
In a multi sensor setting there is spatial context as well.
Table 1: Generalization accuracies of different experiments
experiment
synthetic
left vs. right motor
auditory vs. navigation
L '
'
'=
conditioning
L' = ( ' > : :- ) ' (> ' >0> = )
< ' ( ' = > ) ' (> ' > :)
' > '> =
' (= ' > )
marginalize features
full integration
evaluations show that theoretical arguments are confirmed by significant improvements in
generalization accuracy. The only disadvantage of having a latent feature space is that
all computations get more involved, since there are additional variables that have to be
integrated over. However this additional complexity does not render the method intractable
since the algorithm remains polynomial in the number of segments to be classified. Finally
we want to point out that the improvements observed in our results can only be attributed
to the idea of using a latent feature space. This idea is certainly not limited to time series
classification and should generalize well to other hierarchical architectures.
Acknowledgments
We want to express gratitude to Dr. Rezek, who made several valuable suggestions in the
early stages of this work. We also want to thank Prof. Stokes, who provided us with the
EEG recordings that were used in the experiments section. Finally we are also grateful for
the valuable comments provided by the reviewers of this paper. Peter Sykacek is currently
funded by grant Nr. F46/399 kindly provided by the BUPA foundation.
References
[BS94] J. M. Bernardo and A. F. M. Smith. Bayesian Theory. Wiley, Chichester, 1994.
[GG84] S. Geman and D. Geman. Stochastic relaxation, Gibbs distributions and the Bayesian
restoration of images. IEEE Transactions on Pattern Analysis and Machine Intelligence,
6:721?741, 1984.
[Gre95] P. J. Green. Reversible jump Markov chain Monte Carlo computation and Bayesian model
determination. Biometrika, 82:711?732, 1995.
[Mac92] D. J. C. MacKay. The evidence framework applied to classification networks. Neural
Computation, 4:720?736, 1992.
[Nea96] R. M. Neal. Bayesian Learning for Neural Networks. Springer, New York, 1996.
? Ruanaidh and W. J. Fitzgerald. Numerical Bayesian Methods Applied to Signal
[RF95] J. J. K. O
Processing. Springer-Verlag, New York, 1995.
[RG97] S. Richardson and P. J. Green. On Bayesian analysis of mixtures with an unknown number
of components. Journal Royal Stat. Soc. B, 59:731?792, 1997.
[RJ86] L. R. Rabiner and B. H. Juang. An introduction to Hidden Markov Models. IEEE ASSP
Magazine, 3(1):4?16, 1986.
[RL96] A. E. Raftery and S. M. Lewis. Implementing MCMC. In W.R. Gilks, S. Richardson, and
D.J. Spiegelhalter, editors, Markov Chain Monte Carlo in practice, chapter 7, pages 115?
130. Chapman & Hall, London, Weinheim, New York, 1996.
[Rob96] C. P. Robert. Mixtures of distributions: inference and estimation. In W. R. Gilks,
S. Richardson, and D.J. Spiegelhalter, editors, Markov Chain Mont Carlo in Practice, pages
441?464. Chapman & Hall, London, 1996.
[Syk00] P. Sykacek. On input selection with reversible jump Markov chain Monte Carlo sampling.
In S.A. Solla, T.K. Leen, and K.-R. M?uller, editors, Advances in Neural Information Processing Systems 12, pages 638?644, Boston, MA, 2000. MIT Press.
| 2096 |@word ruanaidh:1 determinant:1 version:1 polynomial:1 stronger:1 suitably:1 covariance:7 necessity:1 series:13 denoting:2 comparing:1 analysed:1 numerical:1 realistic:1 shape:1 motor:7 update:7 v:2 stationary:1 generative:2 intelligence:1 iso:1 smith:1 short:1 regressive:2 node:5 successive:3 direct:1 become:1 pairing:1 qualitative:1 consists:1 combine:1 manner:1 multi:1 weinheim:1 provided:6 notation:1 what:1 kind:1 eigenvector:1 proposing:1 unobserved:1 transformation:1 certainty:2 esti:1 every:2 temporal:1 act:1 bernardo:1 biometrika:1 classifier:3 k2:2 uk:4 grant:1 omit:1 engineering:2 consequence:1 oxford:4 subscript:1 dynamically:1 suggests:1 bupa:1 limited:1 range:2 statistically:1 directed:2 acknowledgment:1 gilks:2 testing:2 recursive:1 practice:2 implement:1 area:2 empirical:1 significantly:1 convenient:3 integrating:4 suggest:1 get:8 marginalize:1 selection:1 context:2 applying:3 seminal:1 dam:1 conventional:2 reviewer:1 resolution:1 updated:2 target:2 suppose:1 magazine:1 us:5 trick:1 element:1 updating:2 geman:2 observed:5 calculate:1 solla:1 valuable:2 mentioned:1 complexity:1 fitzgerald:1 dynamic:1 trained:1 depend:1 grateful:1 segment:13 resolved:1 joint:1 represented:2 chapter:1 derivation:1 london:2 monte:4 hyper:1 lag:2 larger:3 statistic:5 richardson:3 transform:1 obviously:1 advantage:2 sequence:5 eigenvalue:1 propose:3 product:1 fea:1 p4:1 convergence:1 electrode:2 requirement:1 juang:1 depending:1 ac:2 stat:1 soc:1 implemented:1 convention:1 differ:1 concentrate:1 bilistic:1 closely:1 filter:1 stochastic:1 implementing:1 generalization:6 summation:1 hall:2 ground:1 driving:1 early:1 estimation:1 label:11 currently:1 healthy:1 uller:1 mit:1 sensor:7 gaussian:12 avoid:1 probabilistically:1 improvement:4 likelihood:15 contrast:1 nea96:2 sense:1 inference:8 dependent:1 integrated:2 entire:1 typically:1 hidden:12 relation:1 classification:11 proposes:1 spatial:2 integration:4 mackay:1 marginal:1 once:1 extraction:6 having:1 sampling:5 chapman:2 represents:5 look:1 filling:1 gamma:2 individual:1 attempt:2 interest:5 acceptance:2 circular:1 highly:1 evaluation:2 certainly:1 chichester:1 mixture:4 bracket:2 navigation:1 unconditional:1 hg:1 chain:5 integral:1 partial:2 necessary:1 theoretical:1 increased:1 classify:2 ar:14 disadvantage:1 restoration:1 maximization:1 sjrob:1 dependency:2 synthetic:5 combined:1 density:1 probabilistic:5 together:1 fused:1 again:7 recorded:3 dr:1 cognitive:4 imagination:3 summarized:1 student:3 coefficient:23 depends:1 performed:2 observing:4 doing:1 mastoid:1 bayes:1 ass:1 square:2 accuracy:4 variance:3 who:2 correspond:1 rabiner:1 generalize:1 bayesian:12 carlo:5 confirmed:1 published:1 classified:1 explain:1 involved:1 naturally:1 attributed:1 gain:1 gg84:2 auditory:2 sampled:1 knowledge:1 higher:1 follow:1 improved:1 leen:1 done:3 ox:2 furthermore:1 just:1 sykacek:3 stage:6 correlation:2 hastings:1 ox1:2 nonlinear:1 reversible:4 autocorrelations:1 artifact:1 indicated:1 building:1 effect:4 normalized:1 hence:9 neal:1 white:1 adjacent:2 during:3 nuisance:1 recurrence:1 evident:1 reflection:5 mac92:2 pro:1 image:3 wise:1 novel:2 recently:1 multinomial:7 functional:1 conditioning:4 discussed:1 interpretation:2 significant:3 gibbs:3 dag:6 funded:1 robot:2 stable:1 afor:1 posterior:1 own:1 multivariate:1 verlag:1 seen:1 additional:2 determine:1 signal:4 stephen:1 levinson:1 full:6 upside:1 reduces:1 determination:1 cross:1 prediction:2 regression:1 expectation:1 represent:4 kernel:1 invert:1 proposal:5 whereas:4 conditionals:2 separately:1 want:3 comment:1 subject:2 recording:2 hz:3 contrary:1 call:1 distributions1:1 ture:1 variate:1 architecture:4 suboptimal:1 idea:3 whether:1 motivated:1 render:1 peter:2 speaking:1 york:3 repeatedly:1 band:2 generate:2 exist:1 estimated:1 discrete:2 hyperparameter:1 express:1 rezek:1 pj:2 ce:1 fuse:1 graph:3 relaxation:1 sum:1 run:2 inverse:3 fourth:1 reasonable:1 decision:1 bit:1 fold:1 flat:1 encodes:1 argument:3 department:2 according:2 conjugate:1 across:1 metropolis:1 modification:1 jeffreys:1 taken:1 equation:7 previously:1 remains:1 turn:1 count:2 know:1 tractable:1 adopted:1 generalizes:1 gaussians:2 operation:1 apply:3 observe:2 hierarchical:6 denotes:5 assumes:1 dirichlet:4 top:2 remaining:1 prof:1 establish:1 classical:1 move:2 quantity:6 diagonal:4 nr:1 link:3 thank:1 lateral:1 hmm:2 assuming:1 modeled:1 index:1 illustration:1 gji:1 ratio:2 setup:1 robert:2 sharper:1 relate:1 unknown:2 perform:2 observation:14 reparameterise:2 markov:14 mate:1 finite:1 parietal:1 situation:1 stokes:1 assp:1 ofb:1 gratitude:1 c3:2 c4:2 suggested:1 below:4 pattern:1 mismatch:1 built:2 green:2 royal:1 indicator:4 residual:2 improve:1 spiegelhalter:2 raftery:1 auto:4 extract:1 prior:23 understanding:1 expect:4 suggestion:1 acyclic:2 validation:1 foundation:1 integrate:2 sufficient:5 consistent:1 editor:3 summary:1 placed:1 mont:1 midpoint:1 distributed:2 regard:4 dimension:1 transition:4 autoregressive:1 made:1 jump:4 preprocessing:3 transaction:1 conceptual:1 continuous:1 latent:21 table:3 eeg:3 untrained:1 tween:1 significance:1 kindly:1 noise:7 child:2 site:1 wiley:1 precision:5 third:4 jacobian:1 young:1 e4:2 down:1 symbol:1 evidence:1 intractable:1 illustrates:1 t4:1 boston:1 likely:1 springer:2 corresponds:1 lewis:1 ma:1 conditional:7 replace:1 discriminate:1 pas:2 arises:1 meant:1 mcmc:3 correlated:2 |
1,203 | 2,097 | Active Learning
in the Drug Discovery Process
Manfred K. Warmuth , Gunnar
R?atsch , Michael
Mathieson ,
Jun Liao , Christian Lemmen
Computer
Science Dep., Univ. of Calif. at Santa Cruz
FHG
FIRST,
Kekul?estr. 7, Berlin, Germany
DuPont Pharmaceuticals,150 California St. San Francisco.
manfred,mathiesm,liaojun @cse.ucsc.edu, [email protected],
[email protected]
Abstract
We investigate the following data mining problem from Computational
Chemistry: From a large data set of compounds, find those that bind to
a target molecule in as few iterations of biological testing as possible. In
each iteration a comparatively small batch of compounds is screened for
binding to the target. We apply active learning techniques for selecting
the successive batches.
One selection strategy picks unlabeled examples closest to the maximum
margin hyperplane. Another produces many weight vectors by running
perceptrons over multiple permutations of the data. Each weight vector
votes with its prediction and we pick the unlabeled examples for which
the prediction is most evenly split between and . For a third selection strategy note that each unlabeled example bisects the version space
of consistent weight vectors. We estimate the volume on both sides of
the split by bouncing a billiard through the version space and select unlabeled examples that cause the most even split of the version space.
We demonstrate that on two data sets provided by DuPont Pharmaceuticals that all three selection strategies perform comparably well and are
much better than selecting random batches for testing.
1 Introduction
Two of the most important goals in Computational Drug Design are to find active compounds in large databases quickly and (usually along the way) to obtain an interpretable
model for what makes a specific subset of compounds active. Activity is typically defined
All but last author received partial support from NSF grant CCR 9821087
Current address: Austrialian National University, Canberra, Austrialia. Partially supported by
DFG
(JA 379/9-1, MU 987/1-1) and travel grants from EU (Neurocolt II).
Current address: BioSolveIT GmbH, An der Ziegelei 75, Sankt Augustin, Germany
as binding to a target molecule. Most of the time an iterative approach to the problem is
employed. That is in each iteration a batch of unlabeled compounds is screened against the
target using some sort of biological assay[MGST97]. The desired goal is that many active
hits show up in the assays of the selected batches.
From the Machine Learning point of view all examples (compounds) are initially unlabeled. In each iteration the learner selects a batch of un-labeled examples for being labeled
as positive (active) or negative (inactive). In Machine Learning this type of problem has
been called ?query learning? [Ang88] ?selective sampling? [CAL90] or ?active learning?
[TK00]. A Round0 data set contains 1,316 chemically diverse examples, only 39 of which
are positive. A second Round1 data set has 634 examples with 150 positives. 1 This data
set is preselected on the basis of medicinal chemistry intuition. Note that our classification problem is fundamentally asymmetric in that the data sets have typically many more
negative examples and the Chemists are more interested in the positive hits because these
compounds might lead to new drugs. What makes this problem challenging is that each
compound is described by a vector of 139,351 binary shape features. The vectors are
sparse (on the average 1378 features are set per Round0 compound and 7613 per Round1
compound).
We are working with retrospective data sets for which we know all the labels. However, we
simulate the real-life situation by initially hiding all labels and only giving to the algorithm
the labels for the requested batches of examples (virtual screening). The long-term goal of
this type of research is to provide a computer program to the Chemists which will do the
following interactive job: At any point new unlabeled examples may be added. Whenever
a test is completed, the labels are given to the program. Whenever a new test needs to be set up, the
Chemist asks the program to suggest a batch of unlabeled compounds. The suggested batch might be
?edited? and augmented using the invaluable knowledge and intuition of the medicinal Chemist. The
hope is that the computer assisted approach allows
for mining larger data sets more quickly. Note that
compounds are often generated with virtual Combinatorial Chemistry. Even though compound descriptors can be computed, the compounds have not been
Figure 1: Three types of comsynthesized yet. In other words it is comparatively pounds/points:
are active,
are
easy to generate lots of unlabeled data.
inactive and
are yet unlabeled. The
Maximum Margin Hyperplane is used as
In our case the Round0 data set consists of compounds from Vendor catalog and corporate collec- the internal classifier.
tions. Much more design effort went into the harder Round1 data set. Our initial results are
very encouraging. Our selection strategies do much better than choosing random batches
indicating that the long-term goal outlined above may be feasible.
Thus from the Machine Learning point of view we have a fixed set of points in
that
are either unlabeled or labeled positive or negative. (See Figure 1). The binary descriptors
of the compounds are rather ?complete? and the data is always linearly separable. Thus
we concentrate on simple linear classifiers in this paper. 2 We analyzed a large number
of different ways to produce hyperplanes and combine hyperplanes. In the next section we
describe different selection strategies on the basis of these hyperplanes in detail and provide
an experimental comparison. Finally in Section 3 we give some theoretical justification for
why the strategies are so effective.
1
2
Data provided by DuPont Pharmaceuticals.
On the current data sets kernels did not improve the results (not shown).
2 Different Selection Criteria and their Performance
A selection algorithm is specified in three parts: a batch size, an initialization and a selection strategy. In practice it is not cost effective to test single examples at a time. We
always chose 5% of the total data set as our batch size, which matches reasonably with
typical experimental constraints. The initial batches are chosen at random until at least one
positive and one negative example are found. Typically this is achieved with the first batch.
All further batches are chosen using the selection strategy.
As we mentioned in the introduction, all our selection strategies are based on linear classifiers of the data labeled so far.
examples
All
are normalized to unit-length and we consider
homogeneous hyperplanes
is again unit
where the normal direction
length. A plane predicts with sign
on the example/compound .
Once we specify how the weight vector is found then the next batch is found by selecting the
unlabeled examples closest to this hyperplane. The simplest way to obtain such a weight
vector is to run a perceptron over the labeled data until it produces a consistent weight
vector (Perc). Our second selection strategy (called SVM) uses the maximum margin hyperplane [BGV92] produced by a Support Vector Machine. When using the perceptron
to predict for example handwritten characters, it has been shown that ?voting? the predictions of many hyperplanes improves the predictive performance [FS98]. So we always
start from the weight vector zero and do multiple passes over the data until the perceptron
is consistent. After processing each example we store the weight vector. We remember
all weight vectors for each pass 3 and do this for 100 random permutations of the labeled
examples. Each weight vector gets one vote. The prediction on an example is positive if
the total vote is larger than zero and we select the unlabeled examples whose total vote is
closest to zero4 . We call this selection strategy VoPerc.
The dot product is commutative. So when
then the point lies on the positive side
of the hyperplane . In a dual view the point lies on the positive side of the hyperplane
(Recall all instances and weight vectors
that
have
unit-length). A
weight vector
is
must lie on the -side of the plane
for
consistent with all -labeled examples
all . The set of all consistent weight vectors is called the version space which is a section
of the unit hypersphere bounded by the planes corresponding to the labeled examples. An
unlabeled hyperplane
bisects the version space. For our third selection strategy
(VolEst)
of
bounce
a billiard is bounced 1000 times
inside
the
version
space
and
the
fraction
!
"
#
points on the positive side of
is computed. The prediction for
is positive if ! is
larger than half and the strategy selects unlabeled points whose fraction is closest to half.
In Figure 2 (left) we plot the true positives and false positives w.r.t. the whole data set
for Perc and VoPerc showing that VoPerc performs slightly better. Also VoPerc has lower
variance (Figure 2 (right)). Figure 3 (left) shows the averaged true positives and false
positives of VoPerc, SVM, and VolEst. We note that all three perform similarly. We also
plotted ROC curves after each batch has been added (not shown). These plots also show
that all three strategies are comparable.
The three strategies VoPerc, SVM, and VolEst all perform much better than the corresponding strategies where the selection criterion is to select random unlabeled examples instead
of using a ?closest? criterion. For example we show in Figure 4 that SVM is significantly
better than SVM-Rand. Surprisingly the improvement is larger on the easier Round0 data
set. The reason is that the Round0 has a smaller fraction of positive examples (3%). Recall
3
Surprisingly with some smart bookkeeping this can all be done with essentially no computational
overhead. [FS98]
4
Instead of voting the predictions of all weight vectors one can also average all the weight vectors
after normalizing them and select unlabeled examples closest to the resulting single weight vector.
This way of averaging leads to slightly worse results (not shown).
30
25
standard deviation
number of examples
150
100
50
0
0
Perc true pos
Perc false pos
VoPerc true pos
VoPerc false pos
0.2
0.4
0.6
0.8
fraction of examples selected
Perc true pos
Perc false pos
VoPerc true pos
VoPerc false pos
20
15
10
5
0
0
1
0.2
0.4
0.6
0.8
fraction of examples selected
1
Figure 2: (left) Average (over 10 runs) of true positives and false positives on the entire Round1 data
set after each 5% batch for Perc and VoPerc. (right) Standard deviation over 10 runs.
150
total number of hits
number of examples
150
100
50
0
0
VoPerc true pos
VoPerc false pos
SVM true pos
SVM false pos
VolEst true pos
VolEst false pos
0.2
0.4
0.6
0.8
fraction of examples selected
100
50
5% batch size
1 example batch size
1
0
0
0.2
0.4
0.6
0.8
fraction of examples selected
1
Figure 3: (left) Average (over 10 runs) of true and false positives on entire Round1 data set after each
5% batch for VoPerc, SVM, and VolEst. (right) Comparison of 5% batch size and 1 example batch
size for VoPerc on Round1 data.
that the Round1 data was preselected by the Chemists for actives and the fraction was raised
to about 25%. This suggest that our methods are particularly suitable when few positive
examples are hidden in a large set of negative examples.
The simple strategy SVM of choosing unlabeled examples closest to the maximum margin
hyperplane has been investigated by other authors (in [CCS00] for character recognition
and in [TK00] for text categorization). The labeled points that are closest to the hyperplane
are called the support vectors because if all other points are removed then the maximum
margin hyperplane remains unchanged. In Figure 5 we visualize the location of the points
in relation to the center of the hyperplane. We show the location of the points projected
onto the normal direction of the hyperplane. For each 5% batch the location of the points
is scattered onto a thin stripe. The hyperplane crosses the stripe in the middle. In the left
plot the distances are scaled so that the support vectors are at distance 1. In the right
plot the geometric distance to the hyperplane is plotted. Recall that we pick unlabeled
points closest to the hyperplane (center of the stripe). As soon as the ?window? between
the support vectors is cleaned most positive examples have been found (compare with the
SVM curves given in Figure 3 (left)). Also shrinking the width of the geometric window
corresponds to improved generalization.
So far our three selection strategies VoPerc, SVM and VolEst have shown similar performance. The question is whether the performance criterion considered so far is suitable
for the drug design application. Here the goal is to label/verify many positive compounds
quickly. We therefore think that the total number of positives (hits) among all examples
tested so far is the best performance criterion. Note that the total number of hits of the random selection strategy grows linearly with the number of batches (In each random batch
40
150
30
number of examples
number of examples
35
25
20
random true pos
random false pos
closest true pos
closest false pos
15
10
100
random true pos
random false pos
closest true pos
closest false pos
50
5
0
0
0.2
0.4
0.6
0.8
fraction of examples selected
0
0
1
0.2
0.4
0.6
0.8
fraction of examples selected
1
1
0.95
0.9
0.85
0.8
0.75
0.7
0.65
0.6
0.55
0.5
0.45
0.4
0.35
0.3
0.25
0.2
0.15
0.1
0.05
0
?2
1
0.95
0.9
0.85
0.8
0.75
0.7
0.65
0.6
0.55
0.5
0.45
0.4
0.35
0.3
0.25
0.2
0.15
0.1
0.05
0
?0.3
fraction of examples selected
fraction of examples selected
Figure 4: Comparisons of SVM using random batch selection and closest batch selection. (left)
Round0 data. (right) Round1 data.
?1.5
?1
?0.5
0
0.5
1
1.5
2
?0.2
normalized distance to hyperplane
?0.1
0
0.1
0.2
0.3
geometric distance to hyperplane
Figure 5: (left) Scatter plot of the distance of examples to the maximum margin hyperplane normalized so support vectors are at 1. (right) Scatter plot of the geometric distance of examples to the
hyperplane. Each stripe shows location of a random sub-sample of points (Round1 data) after an
additional 5% batch has been labeled by SVM. Selected examples are black x, unselected positives
are red plus, unselected negatives are blue square.
we expect 5% hits). In contrast the total number of hits of VoPerc, SVM and VolEst is
5% in the first batch (since it is random) but much faster thereafter (See Figure 6). VoPerc
performs the best.
Since the positive examples are much more valuable in our application, we also changed
5
the
examples of largest positive distance
selection strategy SVM to selecting unlabeled
to the maximum margin hyperplane (SVM ) rather than smallest distance
.
Correspondingly VoPerc picks the unlabeled example with the highest vote and VolEst
picks the unlabeled example with the largest fraction ! . The total hit plots of the resulting
modified strategies SVM , VoPerc and VolEst are improved ( see Figure 7 versus Figure
6 ). However the generalization plots of the modified strategies (i.e. curves like Figure
3(left)) are slightly worse for the new versions. Thus in some sense the original strategies
are better at ?exploration? (giving better generalization on the entire data set) while the
modified strategies are better at ?exploitation? (higher number of total hits). We show this
trade-off in Figure
8 for SVM and SVM . The same trade-off occurs for the VoPerc
and VolEst
pairs of strategies(not shown).
Finally we investigate the effect of batch size on performance. For simplicity we only show
total hit plots for VoPerc( Figure 3 (right) ). Note that for our data a batch size of 5% (31
examples for Round1) is performing not much worse than the experimentally unrealistic
batch size of only 1 example. Only when the results for batch size 1 are much better than
5
In Figure 5 this means we are selecting from right to left
40
150
30
total number of hits
total number of hits
35
25
20
15
10
5
0
0
VoPerc
SVM
VolEst
0.2
0.4
0.6
0.8
fraction of examples selected
50
0
0
1
VoPerc
SVM
VolEst
100
0.2
0.4
0.6
0.8
fraction of examples selected
Figure 6: Total hit performance on Round0 (left) and Round1 (right) data of
VolEst with 5% batch size.
40
1
, VoPerc and
150
30
total number of hits
total number of hits
35
25
20
15
10
5
0
0
VoPerc+
SVM+
VolEst+
0.2
0.4
0.6
0.8
fraction of examples selected
1
100
50
VoPerc+
SVM+
VolEst+
0
0
0.2
0.4
0.6
0.8
fraction of examples selected
Figure 7: Total hit performance on Round0 (left) and Round1 (right) data of
VolEst with 5% batch size.
1
, VoPerc and
the results for larger batch sizes, more sophisticated selection strategies are worth exploring
that pick say a batch that is ?close? and at the same time ?diverse?.
At this point our data sets are still small enough that we were able to precompute all dot
products
(the kernel matrix). After this preprocessing, one pass of a perceptron is at most
, where is the number of labeled examples and
number of mistakes.
the
Finding the maximum margin hyperplane
is estimated at
time. For the computa
tion of VolEst we need to spend
per bounce of the billiard. In our implementations
we used SVM Light [Joa99] and the billiard algorithm of [Ruj97, RM00, HGC99].
If we have the internal hypothesis of the algorithm then for applying the selection criterion
we need to evaluate the hypothesis for each unlabeled point. This cost is proportional to
the number of support vectors for the SVM-based methods and proportional to the number
of mistakes for the perceptron-based methods. In the case of VolEst we again need
time per bounce, where is the number of labeled points.
Overall VolEst was clearly the slowest. For much larger data sets VoPerc seems to be the
simplest and the most adaptable.
3 Theoretical Justifications
As we see in Figure 5(right) the geometric margin of the support vectors (half the width
of the window) is shrinking as more examples are labeled. Thus the following goal is reasonable for designing selection strategies: pick unlabeled examples that cause the margin
to shrink the most. The simplest such strategy is to pick examples closest to the maximum margin hyperplane since these example are expected to change the maximum margin
150
number of examples
total number of hits
150
100
50
0
0
SVM+
SVM
0.2
0.4
0.6
0.8
1
fraction of examples selected
100
SVM+ true pos
SVM true pos
SVM+ false pos
SVM false pos
50
0
0
0.2
0.4
0.6
0.8
1
fraction of examples selected
Figure 8: Exploitation versus Exploration: (left) Total hit performance and (right) True and False
positives performance (right) of SVM and on Round 1 data
hyperplane the most [TK00, CCS00].
An alternative goal is to reduce the volume of the version space. This volume is a rough
measure of the remaining uncertainty
and
in the data. Recall that both the weight vectors
instances have unit length. Thus
is the distance
of the point to the plane as well
to the plane . The maximum margin
as (in the dual
view) the distance
of the point
hyperplane
is the point
in version space with the largest sphere that is completely
contained in the version space [Ruj97, RM00]. After labeling only one side of the plane
remains. So if passes close to the point
then about half of the largest sphere is
eliminated from the version space. So this is a second justification for selecting unlabeled
examples closest to the maximum margin hyperplane.
Our selection strategy VolEst starts from any point inside the version space and then
bounces a billiard 1000 times.
The billiard is almost always ergodic (See discussion in
[Ruj97]).
Thus the fraction ! of bounces on the positive side of an unlabeled hyperplane
is an
. Since it is unknown
estimate of the fraction of volume on the positive side of
how
will be labeled, the best example are those
that
split
the
version
space in half. Thus
in VolEst we select unlabeled points for which ! is closest to half. The thinking underlying our strategy VolEst is most closely related to the Committee Machine where random
concepts in the version space are asked to vote on the next random example and the label
of that example is requested only if the vote is close to an even split [SOS92].
We tried to improve our estimate of the volume
by replacing ! by the fraction of the total
#
trajectory located on the positive side of . On our two data sets this did not improve the
performance
(not shown). We also averaged the 1000 bounce points. The resulting weight
vector (an approximation to the center of mass of the version space) approximates the
so called Bayes point [Ruj97] which
has the following property: Any unlabeled hyperplane
passing through the Bayes point cuts the version space roughly 6 in half. We thus tested
a selection strategy which picks unlabeled points closest to the estimated center of mass.
This strategy was again indistinguishable from the other two strategies based on bouncing
the billiard.
We have no rigorous justification for the
variants of our algorithms.
4 Conclusion
We showed how the active learning paradigm ideally fits the drug design cycle. After some
deliberations we concluded that the total number of positive examples (hits) among the
tested examples is the best performance criterion for the drug design application. We found
6
Even in dimension two there is no point that does this exactly [Ruj97].
that a number of different selection strategies with comparable performance. The variants
that select the unlabeled examples with the highest score (i.e. the variants) perform better.
Overall the selection strategies based on the Voted Perceptron were the most versatile and
showed slightly better performance.
References
[Ang88] D. Angluin. Queries and concept learning. Machine Learning, 2:319?342,
1988.
[BGV92] B.E. Boser, I.M. Guyon, and V.N. Vapnik. A training algorithm for optimal
margin classifiers. In D. Haussler, editor, Proceedings of the 5th Annual ACM
Workshop on Computational Learning Theory, pages 144?152, 1992.
[CAL90] D. Cohn, L. Atlas, and R. Ladner. Training connectionist networks with
queries and selective sampling. Advances in Neural Information Processing
Systems, 2:566?573, 1990.
[CCS00] C. Campbell, N. Cristianini, and A. Smola. Query learning with large margin
classifiers. In Proceedings of ICML2000, page 8, Stanford, CA, 2000.
[FS98] Y. Freund and R. Schapire. Large margin classification using the perceptron
algorithm. In Proc. 11th Annu. Conf. on Comput. Learning Theory. ACM
Press, New York, NY, July 1998.
[HGC99] Ralf Herbrich, Thore Graepel, and Colin Campbell. Bayes point machines:
Estimating the bayes point in kernel space. In Proceedings of IJCAI Workshop
Support Vector Machines, pages 23?27, 1999.
[Joa99] T. Joachims. Making large?scale SVM learning practical. In B. Sch?olkopf,
C.J.C. Burges, and A.J. Smola, editors, Advances in Kernel Methods ? Support Vector Learning, pages 169?184, Cambridge, MA, 1999. MIT Press.
[MGST97] P. Myers, J. Greene, J. Saunders, and S. Teig. Rapid, reliable drug discovery.
Today?s Chemist at Work, 6:46?53, 1997.
[RM00] P. Ruj?an and M. Marchand. Computing the bayes kernel classifier. In Advances
in Large Margin Classifiers, volume 12, pages 329?348. MIT Press, 2000.
[Ruj97] P. Ruj?an. Playing billiard in version space. Neural Computation, 9:99?122,
1997.
[SOS92] H. Seung, M. Opper, and H. Sompolinsky. Query by committee. In Proceedings of the Fifth Workshop on Computational Learning Theory, pages 287?
294, 1992.
[TK00] S. Tong and D. Koller. Support vector machine active learning with applications to text classification. In Proceedings of the Seventeenth International
Conference on Machine Learning, San Francisco, CA, 2000. Morgan Kaufmann.
| 2097 |@word bounced:1 exploitation:2 middle:1 version:17 seems:1 cal90:2 tried:1 pick:9 asks:1 versatile:1 harder:1 initial:2 contains:1 score:1 selecting:6 current:3 yet:2 scatter:2 must:1 fs98:3 cruz:1 shape:1 christian:1 dupont:3 plot:9 interpretable:1 atlas:1 half:7 selected:16 warmuth:1 plane:6 manfred:2 hypersphere:1 cse:1 location:4 successive:1 billiard:8 hyperplanes:5 herbrich:1 along:1 ucsc:1 consists:1 combine:1 overhead:1 inside:2 expected:1 rapid:1 roughly:1 encouraging:1 window:3 hiding:1 provided:2 estimating:1 bounded:1 underlying:1 mass:2 what:2 sankt:1 finding:1 remember:1 computa:1 voting:2 interactive:1 exactly:1 classifier:7 hit:19 scaled:1 unit:5 grant:2 positive:31 bind:1 mistake:2 might:2 chose:1 black:1 au:1 initialization:1 plus:1 challenging:1 averaged:2 seventeenth:1 practical:1 testing:2 practice:1 drug:7 significantly:1 word:1 suggest:2 get:1 onto:2 unlabeled:30 selection:26 close:3 applying:1 center:4 ergodic:1 simplicity:1 haussler:1 ralf:1 justification:4 target:4 today:1 ang88:2 homogeneous:1 us:1 designing:1 hypothesis:2 recognition:1 particularly:1 located:1 asymmetric:1 stripe:4 cut:1 predicts:1 database:1 labeled:14 cycle:1 sompolinsky:1 went:1 eu:1 trade:2 removed:1 highest:2 valuable:1 edited:1 mentioned:1 intuition:2 mu:1 asked:1 ideally:1 cristianini:1 seung:1 chemist:6 smart:1 predictive:1 learner:1 basis:2 completely:1 sos92:2 po:26 univ:1 describe:1 effective:2 query:5 labeling:1 choosing:2 saunders:1 whose:2 larger:6 spend:1 stanford:1 say:1 think:1 deliberation:1 myers:1 product:2 olkopf:1 ijcai:1 produce:3 categorization:1 tions:1 dep:1 received:1 job:1 concentrate:1 direction:2 closely:1 exploration:2 virtual:2 ja:1 generalization:3 biological:2 exploring:1 assisted:1 considered:1 normal:2 predict:1 visualize:1 smallest:1 proc:1 travel:1 label:6 combinatorial:1 augustin:1 largest:4 bisects:2 hope:1 rough:1 clearly:1 mit:2 always:4 modified:3 rather:2 joachim:1 improvement:1 slowest:1 contrast:1 rigorous:1 sense:1 typically:3 entire:3 initially:2 hidden:1 relation:1 koller:1 selective:2 fhg:1 selects:2 germany:2 interested:1 overall:2 classification:3 dual:2 among:2 raised:1 once:1 sampling:2 eliminated:1 thin:1 thinking:1 connectionist:1 fundamentally:1 few:2 national:1 pharmaceutical:3 dfg:1 screening:1 investigate:2 mining:2 analyzed:1 light:1 partial:1 calif:1 desired:1 plotted:2 theoretical:2 instance:2 kekul:1 cost:2 deviation:2 subset:1 st:1 international:1 off:2 michael:1 quickly:3 again:3 perc:7 worse:3 conf:1 de:1 chemistry:3 tion:1 view:4 lot:1 red:1 start:2 sort:1 bayes:5 square:1 voted:1 descriptor:2 variance:1 kaufmann:1 handwritten:1 comparably:1 produced:1 trajectory:1 worth:1 whenever:2 against:1 icml2000:1 recall:4 knowledge:1 improves:1 graepel:1 sophisticated:1 campbell:2 adaptable:1 higher:1 specify:1 improved:2 rand:1 done:1 though:1 pound:1 shrink:1 smola:2 until:3 working:1 replacing:1 cohn:1 grows:1 thore:1 effect:1 normalized:3 true:18 verify:1 concept:2 assay:2 round:1 indistinguishable:1 width:2 criterion:7 complete:1 demonstrate:1 performs:2 invaluable:1 estr:1 bookkeeping:1 volume:6 approximates:1 raetsch:1 cambridge:1 outlined:1 similarly:1 dot:2 closest:18 showed:2 compound:18 store:1 binary:2 life:1 der:1 morgan:1 additional:1 employed:1 paradigm:1 colin:1 july:1 ii:1 multiple:2 corporate:1 match:1 faster:1 cross:1 long:2 sphere:2 prediction:6 variant:3 liao:1 essentially:1 iteration:4 kernel:5 achieved:1 concluded:1 sch:1 pass:2 call:1 split:5 easy:1 enough:1 fit:1 reduce:1 bounce:6 inactive:2 whether:1 effort:1 retrospective:1 passing:1 cause:2 york:1 santa:1 mathieson:1 simplest:3 generate:1 angluin:1 schapire:1 nsf:1 sign:1 estimated:2 per:4 ccr:1 blue:1 diverse:2 gunnar:2 thereafter:1 fraction:22 screened:2 run:4 uncertainty:1 bouncing:2 chemically:1 almost:1 reasonable:1 guyon:1 comparable:2 marchand:1 annual:1 activity:1 greene:1 constraint:1 simulate:1 performing:1 separable:1 precompute:1 smaller:1 slightly:4 character:2 making:1 vendor:1 remains:2 committee:2 know:1 apply:1 batch:38 alternative:1 original:1 running:1 remaining:1 completed:1 giving:2 comparatively:2 unchanged:1 added:2 question:1 occurs:1 strategy:35 distance:11 berlin:1 neurocolt:1 evenly:1 reason:1 ruj:2 length:4 negative:6 design:5 implementation:1 unknown:1 perform:4 ladner:1 situation:1 pair:1 cleaned:1 specified:1 catalog:1 california:1 boser:1 address:2 able:1 suggested:1 usually:1 preselected:2 program:3 reliable:1 unrealistic:1 suitable:2 improve:3 unselected:2 jun:1 text:2 medicinal:2 discovery:2 geometric:5 freund:1 expect:1 permutation:2 proportional:2 versus:2 consistent:5 editor:2 playing:1 changed:1 supported:1 last:1 surprisingly:2 soon:1 side:9 burges:1 perceptron:7 correspondingly:1 fifth:1 sparse:1 curve:3 dimension:1 opper:1 author:2 san:2 projected:1 preprocessing:1 far:4 active:11 francisco:2 un:1 iterative:1 why:1 reasonably:1 molecule:2 ca:2 requested:2 investigated:1 did:2 linearly:2 whole:1 gmbh:1 augmented:1 canberra:1 roc:1 scattered:1 ny:1 tong:1 shrinking:2 sub:1 comput:1 lie:3 lemmen:1 third:2 annu:1 specific:1 showing:1 svm:33 normalizing:1 workshop:3 false:18 vapnik:1 commutative:1 anu:1 margin:18 easier:1 contained:1 partially:1 binding:2 corresponds:1 acm:2 ma:1 goal:7 feasible:1 experimentally:1 change:1 typical:1 hyperplane:27 averaging:1 called:5 total:20 pas:2 experimental:2 vote:7 atsch:1 perceptrons:1 select:6 indicating:1 internal:2 support:11 evaluate:1 tested:3 |
1,204 | 2,098 | Transform-invariant image decomposition
with similarity templates
Chris Stauffer, Erik Miller, and Kinh Tieu
MIT Artificial Intelligence Lab
Massachusetts Institute of Technology
Cambridge, MA 02139
{stauffer,emiller,tieu}@ai.mit.edu
Abstract
Recent work has shown impressive transform-invariant modeling
and clustering for sets of images of objects with similar appearance.
We seek to expand these capabilities to sets of images of an object
class that show considerable variation across individual instances
(e.g. pedestrian images) using a representation based on pixel-wise
similarities, similarity templates. Because of its invariance to the
colors of particular components of an object, this representation enables detection of instances of an object class and enables alignment
of those instances. Further, this model implicitly represents the regions of color regularity in the class-specific image set enabling a
decomposition of that object class into component regions.
1
Introduction
Images of a class of objects are often not effectively characterized by a Gaussian
distribution or even a mixture of Gaussians. In particular, we are interested in
modeling classes of objects that are characterized by similarities and differences
between image pixels rather than by the values of those pixels. For instance, images
of pedestrians (at a certain scale and pose) can be characterized by a few regions
of regularity (RORs) such as shirt, pants, background, and head, that have fixed
properties such as constant color or constant texture within the region, but tend to
be different from each other. The particular color (or texture) of those regions is
largely irrelevant. We shall refer to sets of images that fit this general description
as images characterized by regions of regularity, or ICRORs.
Jojic and Frey [1] and others [2] have investigated transform-invariant modeling
and clustering for images of a particular object (e.g., an individual?s face). Their
method can simultaneously converge on a model and align the data to that model.
This method has shown positive results for many types of objects that are effectively
modeled by a Gaussian or a mixture of Gaussians. Their work with transformed
component analysis (TCA) shows promise for handling considerable variation within
the images resulting from lighting or slight misalignments. However, because these
models rely on an image set with a fixed mean or mixture of means, they are not
directly applicable to ICRORs.
We would also like to address transform-invariant modeling, but use a model which
is invariant to the particular color of component regions. One simple way to achieve
this is to use edge templates to model local differences in image color. In contrast,
we have chosen to model global similarities in color using a similarity template (ST).
While representations of pixel similarity have previously been exploited for segmentation of single images [3, 4], we have chosen to use them for aggregate modeling
of image sets. Similarity templates enable alignment of image sets and decomposition of images into class-specific pixel regions. We note also that registration of
two ICRORs can be accomplished by minimizing the mutual information between
corresponding pixels [5]. But, there is no obvious way of extending this method to
large sets of images without a combinatorial explosion.
Section 2 briefly introduces similarity templates. We investigate their uses for modeling and detection. Section 3 discusses dataset alignment. Section 4 covers their
application to decomposing a class-specific set of images into component regions.
Future avenues of research and conclusions are discussed Section 5.
2
Similarity templates
This section begins with a brief explanation of the similarity template followed
by the mechanics of computing and comparing similarity templates. A similarity
template S for an N -pixel image is an N xN matrix. The element Si,j represents
the probability that pixel locations pi and pj would result from choosing a region
and drawing (iid) two samples (pixel locations) from it. More formally,
X
Si,j =
p(r)p(pi |r)p(pj |r),
(1)
r
where p(r) is the probability of choosing region r and p(pi |r) is the probability of
choosing pixel location pi from region r.
2.1
The ?ideal? similarity template
Consider sampling pixel pairs as described above from an N -pixel image of a particular object (e.g., a pedestrian) segmented by an oracle into disjoint regions (e.g.,
shirt, pants, head, feet, background). Assuming each region is equally likely to be
sampled and that the pixels in the region are selected with uniform probability,
then
1 1 2
( R )( Sr ) if ri = rj
Si,j =
(2)
0
otherwise,
where R is the number of regions, Sr is the number of pixels in region r, and ri
is the region label of pi . If two pixels are from the same region, the corresponding
value is the product of the probability R1 of choosing a particular region and the
probability ( S1r )2 of drawing that pixel pair. This can be interpreted as a block
diagonal co-occurrence matrix of sampled pixel pairs.
In this ideal case, two images of different pedestrians with the same body size and
shape would result in the same similarity template regardless of the colors of their
clothes, since the ST is a function only of the segmentation. An ST of an image
without a pedestrian would exhibit different statistics. Note that even the ST of
an image of a blank wall (segmented as a single region) would be different because
pixels that are in different regions under the ideal pedestrian ST would be in the
same region.
Unfortunately, images do not typically come with labeled regions, and so computation of a similarity template is impossible. However, in this paper, we take
advantage of the observation that properties within a region, such as color, are
often approximately constant. Using this observation, we can approximate true
similarity templates from unsegmented images.
2.2
Computing similarity templates
For the purposes of this paper, our model for similarity is based solely on color.
Since there is a correlation between color similarity and two pixels being in the
same region, we approximate the corresponding value S?i,j with a measure of color
similarity:
1
?||Ii ? Ij ||2
?
Si,j =
,
(3)
exp
N Zi
?i2
where Ii and Ij are pixel color values, ?i2 is a parameter that adjusts the color
similarity measure as a function of the pixel color distribution in the image, and Zi
is the sum of the ith row. This normalization is required because large regions have
a disproportionate effect on the ST estimate. The choice of ?i2 had little effect on
the resulting ST.
If each latent region had a constant but unique color and the regions were of equal
size, then as ?i2 approaches zero this process reconstructs the ?ideal? similarity
template defined in Equation 1. Although region colors are neither constant nor
unique, this approximation has proven to work well in practice.
It is possible to add a spatial prior based on the relative pixel location to model the
fact that similarities tend to local, but we will rely on the statistics of the images
in our data set to determine whether (and to what extent) this is the case. Also, it
may be possible to achieve better results using a more complex color model (e.g.,
hsv with full covariance) or broadening the measure of similarity to include other
modalities (e.g., texture, motion, depth, etc.).
Figure 1 shows two views of the same similarity template. The first view represents
each pixel?s similarity to every other pixel. The second view contains a sub-image
for each pixel which highlights the pixels that are most likely produced by the same
region. Pixels in the shirt tend to highlight the entire shirt and the pants (to a
lesser amount). Pixels in the background tend to be very dissimilar to all pixels in
the foreground.
2.3
Aggregate similarity templates (AST)
We assume each estimated ST is a noisy measurement of the true underlying joint
distribution. Hence we compute an aggregate similarity template (AST) as the
mean S? of the ST estimates over an entire class-specific set of K images:
K
1 X ?k
Si,j .
S?i,j =
K
(4)
k=1
For this quantity to be meaningful, the RORs must be in at least partial correspondence across the training set. Note that this is a less restrictive assumption than
assuming edges of regions are in correspondence across an image set, since regions
have greater support. Being the mean of a set of probability distributions, the AST
is also a valid joint probability distribution.
(a)
(b)
Figure 1: (a) The N xN aggregate similarity template for pedestrian data set. (b)
An alternate view of (a). This view is a width2 xheight2 version of (a). Each subimage represents the row of the original AST that corresponds to that pixel. Each
sub-image highlights the pixels that are most similar to the pixel it represents.
2.4
Comparing similarity templates
To compare an estimated similarity template S? to an aggregate similarity template
S? we evaluate their dot product1 :
XX
? S)
? =
s(S,
S?i,j S?i,j .
(5)
i
j
We are currently investigating other measures for comparison. By thresholding the
ratio of the dot product of a particular image patch under and AST trained on
pedestrian image patches versus an AST trained on random image patches, we can
determine whether a person is present in the image. In previous work [6], we have
illustrated encouraging detection performance.
3
Data set alignment
In this paper, we investigate a more difficult problem: alignment of a set of images.
To explore this problem, we created a set of 128x64 images of simulated pedestrians.
These pedestrians were generated by creating four independently-colored regions
corresponding to shirts, pants, head, and background. Each region was given a
random color. The RGB components were chosen from a uniform distribution
[0, 1]. Then, independent Gaussian noise was added to each pixel (? = .1). Finally
the images were translated uniformly up to 25% of the size of the object. Figure 2
shows examples of these images.
1
In our experimentation KL-divergence, typically used to compare estimates of distributions, proved less robust.
Figure 2: A set of randomly generated ?pedestrian? images used in alignment experimetns.
Using the congealing procedure of Miller et al. [2], we iteratively estimated the
latent variables (translations) that maximized the probability of the image STs to
the AST and re-estimated the AST. We were able to align the images to within .5
pixels on average.
4
Decomposing the similarity template
This section explains how to derive a factorized representation from the AST that
will be useful for recognition of particular instances of a class and for further refinement of detection. This representation is also useful in approximating the template
to avoid the O(N 2 ) storage requirements.
An AST represents the similarity of pixels within an image across an entire classspecific data set. Pairwise statistics have been used for segmentation previously
[3]. Recently, work centered on factoring joint distributions has gained increasing
attention [7, 8, 9, 10]. Rather than estimating two sets of marginals (conditioned on
a latent variable) that explain co-occurrence data (e.g. word-document pairs), we
seek a single set of marginals conditioned on a latent variable (the ROR) that explain
our co-occurrence data (pixel position pairs). Hence, it is a density factorization
in which the two conditional factors are identical (Equation 1). We refer to this as
symmetric factorization of a joint density.
Also, rather than treating pixel brightness (darkness, redness, blueness, or hue) as
a value to be reconstructed in the decomposition, we chose to represent pixel similarity. In contrast to simply treating images as additive mixtures of basis functions
[9], our decomposition will get the same results on a database of images of digits
written in black on white paper or in white on a black board and color images
introduce no difficulties for our methods.
We would like to estimate the factors from Equation 1 that best reconstruct our
? Let S? be the estimate of S? constructed from these factors. Given
measured AST, S.
the number of regions R, it is possible to estimate the priors for each region p(r)
and the probability of each region producing each pixel p(pi |r). The error function
we minimize is the KL-divergence between the empirically measured S? and our
?
parameterized estimate S,
!
XX
?i,j
S
E=
(6)
S?i,j log
S?i,j
i
j
as in [8]. Because our model S? is symmetric, this case can be updated with only
two rules:
X
? i , pj )
S(p
pnew (pi |r) ? p(pi |r)
p(r)p(pj |r) ?
, and
(7)
S(pi , pj )
p
j
pnew (r) ? p(r)
XX
pi
pj
? i , pj )
S(p
p(pj |r)p(pi |r) ?
.
S(pi , pj )
(8)
50
100
150
200
250
300
350
400
450
500
50
100
150
200
250
300
350
400
450
500
Figure 3: The similarity template and the corresponding automatically generated
binary decomposition of the images in the pedestrian data set. The root node represents every pixel in the image. The first branch splits foreground vs. background
pixels. Other nodes correspond to shirt, legs, head, and background regions.
The more underlying regions we allow our model, the closer our estimate will approximate the true joint distribution. These region models tend to represent parts
of the object class. p(pi |r) will tend to have high probabilities for a set of pixels
belonging to the same region. We take advantage of the fact that aligned pedestrian
images are symmetric about the vertical axis by adding a ?reflected? aggregate similarity template to the aggregate similarity template. The resulting representation
provides a compact approximation of the AST (O(RN ) rather than O(N 2 )).
Rather than performing a straight R-way decomposition of the AST to obtain R
pixel region models, we extracted a hierarchical segmentation in the form of a binary
tree. Given the initial region-conditioned marginals p(pi |r0 ) and p(pi |r1 ), each pixel
was assigned to the region with higher likelihood. This was iteratively applied to
the ASTs defined for each sub-region. Region priors were set to 0.5 and not adapted
in order to encourage a balanced cut.
The probabilistic segmentation can be employed to accumulate robust estimates of
statistics of the region. For instance, the mean pixel value can be calculated as a
weighted mean where the pixels are weighted by p(pi |r).
4.1
Decomposing pedestrians
Because the data collected at our lab showed limited variability in lighting, background composition, and clothing, we used the MIT CBCL pedestrian data set
which contains images of 924 unique, roughly aligned pedestrians in a wide variety of environments to estimate the AST. Figure 3 shows the resulting hierarchical
segmentation for the pedestrian AST. Since this intuitive representation was derived automatically with absolutely no knowledge about pedestrians, we hope other
classes of objects can be similarly decomposed into RORs.
In our experience, a color histogram of all the pixels within a pedestrian is not
useful for recognition and was almost useless for data mining applications. Here
we propose a class-conditional color model. It determines a color model over each
region that our algorithm has determined contain similar color information within
this class of objects. This allows us to obtain robust estimates of color in the regions
Figure 4: Results of automatic clustering on three components: shirt, pants, and
the background. Each shows the feature, the most unusual examples of that region,
followed by the 12 most likely examples for the eight prototypical colors of that
region.
of regularity. Further, as a result of our probabilistic segmentation, the values of
p(pi |r) indicate which pixels are most regular in a region which enables us to weight
the contribution of each pixel to the color model.
For the case of pedestrian-conditional color models, the regions roughly correspond
to shirt color, pant color, feet color, head color, and some background color regions. The colors in a region of a single image can be modeled by color histograms,
Gaussians, or mixtures of Gaussians. These region models can be clustered across
images to determine a density of shirt colors, pant colors, and other region colors
within a particular environment. This enables not only an efficient factored color
component codebook, but anomaly detection based on particular regions and higher
order models of co-occurrences between particular types of regions. To illustrate
the effectiveness of our representation we chose the simplest model for the colors in
each region?a single Gaussian in RGB space. The mean and variance of each Gaussian was computed by weighting the pixels represented by the corresponding node
by p(pi |r). This biases the estimate towards the ?most similar? pixels in the region
(e.g., the center of the shirt or the center of the legs). This allows us to represent
the colors of each pedestrian image with 31 means and variances corresponding to
the (2treeheight ? 1) nodes.
We investigated unsupervised clustering on components of the conditional color
model. We fit a mixture of eight Gaussians to the 924 color means for each region.
Figure 4 shows the 12 pedestrians with the highest probability under each of the
eight models and the 12 most unusual pedestrians with respect to that region for
three of the nodes of the tree: shirt color, pant color, and color of the background.
Red, white, blue, and black shirts represent a significant portion of the database.
Blue jeans are also very common in the Boston area (where the CBCL database
was collected). Indoor scenes tended to be very dark, and cement is much more
common than grass.
5
Conclusions
While this representation shows promise, it is not ideal for many problems. First,
it is expensive in both memory and computation. Here, we are only using a simple
measure of pairwise similarity?color similarity. In the future, similarity templates
could be applied to different modalities including texture similarity, depth similarity,
or motion similarity.
While computationally intensive, we believe that similarity templates can provide
a unified approach to the extraction of possible class-specific targets from an image
database, alignment of the candidate images, and precomputation of meaningful features of that class. For the case of pedestrians, it could detect potential pedestrians
in a database, align them, derive a model of pedestrians, and extract the parameters
for each pedestrian. Once the features are computed, query and retrieval can be
done efficiently.
We have introduced a new image representation based on pixel-wise similarity. We
have shown its application in both alignment and decomposition of pedestrian images.
References
[1] Jojic, N. and B. J. Frey. ?Topographic transformation as a discrete latent variable.?
In NIPS 12, S. A. Solla, T. K. Leen and K.-R. Muller (eds), MIT Press, Cambridge,
MA.
[2] Miller, E., N. Matsakis, and P. Viola, (2000) ?Learning from One Example Through
Shared Densities on Transforms.? CVPR2000, Vol. 1, pp. 464-471.
[3] Shi, J. and J. Malik. ?Normalized Cuts and Image Segmentation,? In CVPR San juan,
Puerto Rico, June 1997.
[4] Boykov, Y., O. Veksler and R. Zabih. Fast Approximate Energy Minimization via
Graph Cuts, In ICCV (99), September 1999.
[5] Viola, P. Alignment by Maximization of Mutual Information. MIT Artificial Intelligence Lab, Ph.D. Thesis AI-TR #1548, June, 1995.
[6] Stauffer, C. and W.E.L. Grimson. ?Similarity templates for detection and recognition,?
submitted to CVPR (2001).
[7] Pereira, F.C., N. Tishby, and L. Lee. ?Distributional clustering of English words.?
In 30th Annual Meeting of the Association for Computational Linguistics, Columbus,
Ohio, pages 183?190, 1993.
[8] Thomas Hofmann, ?Probabilistic Latent Semantic Analysis,? UAI (99), Morgan Kaufmann Publishers, Inc., San Francisco, 1999.
[9] Lee, D. D. and H. S. Seung. ?Learning the parts of objects by non-negative matrix
factorization.? Nature 401, 788-791 (1999).
[10] Stauffer, C.. ?Automatic hierarchical classification using time-based co-occurrences.?
CVPR (1999), Fort Colins, CO, (June 1999).
| 2098 |@word version:1 briefly:1 seek:2 rgb:2 decomposition:8 covariance:1 brightness:1 tr:1 initial:1 contains:2 document:1 blank:1 comparing:2 si:5 must:1 written:1 additive:1 hofmann:1 enables:4 shape:1 treating:2 v:1 grass:1 intelligence:2 selected:1 ith:1 colored:1 provides:1 node:5 location:4 hsv:1 codebook:1 kinh:1 constructed:1 introduce:1 pairwise:2 roughly:2 nor:1 mechanic:1 shirt:12 decomposed:1 automatically:2 little:1 encouraging:1 increasing:1 begin:1 xx:3 underlying:2 estimating:1 factorized:1 what:1 interpreted:1 unified:1 clothes:1 transformation:1 every:2 precomputation:1 producing:1 positive:1 frey:2 local:2 solely:1 approximately:1 black:3 chose:2 co:6 factorization:3 limited:1 unique:3 practice:1 block:1 digit:1 procedure:1 area:1 word:2 regular:1 get:1 storage:1 ast:15 impossible:1 darkness:1 center:2 shi:1 regardless:1 attention:1 independently:1 factored:1 adjusts:1 rule:1 x64:1 variation:2 updated:1 target:1 anomaly:1 us:1 element:1 recognition:3 expensive:1 cut:3 distributional:1 labeled:1 database:5 region:64 solla:1 highest:1 balanced:1 grimson:1 environment:2 seung:1 trained:2 ror:1 misalignment:1 tca:1 translated:1 basis:1 joint:5 represented:1 fast:1 artificial:2 query:1 congealing:1 aggregate:7 choosing:4 jean:1 cvpr:3 drawing:2 otherwise:1 reconstruct:1 statistic:4 topographic:1 transform:4 noisy:1 advantage:2 propose:1 product:2 aligned:2 achieve:2 description:1 intuitive:1 regularity:4 requirement:1 extending:1 r1:2 object:15 derive:2 illustrate:1 pose:1 measured:2 ij:2 disproportionate:1 come:1 indicate:1 foot:2 centered:1 enable:1 explains:1 clustered:1 wall:1 clothing:1 exp:1 cbcl:2 purpose:1 applicable:1 combinatorial:1 label:1 currently:1 puerto:1 weighted:2 hope:1 minimization:1 mit:5 gaussian:5 rather:5 avoid:1 derived:1 june:3 likelihood:1 contrast:2 stauffer:4 detect:1 factoring:1 typically:2 entire:3 expand:1 transformed:1 interested:1 pixel:52 classification:1 spatial:1 mutual:2 equal:1 once:1 extraction:1 sampling:1 identical:1 represents:7 unsupervised:1 foreground:2 future:2 others:1 few:1 randomly:1 simultaneously:1 divergence:2 individual:2 classspecific:1 detection:6 investigate:2 mining:1 alignment:9 introduces:1 mixture:6 edge:2 closer:1 explosion:1 partial:1 encourage:1 experience:1 tree:2 re:1 instance:6 modeling:6 cover:1 maximization:1 uniform:2 veksler:1 tishby:1 st:9 person:1 density:4 probabilistic:3 lee:2 thesis:1 reconstructs:1 juan:1 creating:1 potential:1 pedestrian:28 inc:1 sts:1 cement:1 view:5 root:1 lab:3 red:1 portion:1 capability:1 contribution:1 minimize:1 kaufmann:1 variance:2 largely:1 efficiently:1 miller:3 maximized:1 correspond:2 produced:1 iid:1 lighting:2 straight:1 submitted:1 explain:2 tended:1 ed:1 energy:1 pp:1 obvious:1 sampled:2 dataset:1 proved:1 massachusetts:1 color:47 knowledge:1 segmentation:8 rico:1 higher:2 reflected:1 leen:1 done:1 correlation:1 unsegmented:1 columbus:1 believe:1 effect:2 contain:1 true:3 normalized:1 hence:2 assigned:1 jojic:2 symmetric:3 iteratively:2 i2:4 illustrated:1 white:3 semantic:1 motion:2 image:60 wise:2 ohio:1 recently:1 boykov:1 common:2 empirically:1 discussed:1 slight:1 association:1 marginals:3 accumulate:1 refer:2 composition:1 measurement:1 cambridge:2 significant:1 ai:2 automatic:2 similarly:1 had:2 dot:2 similarity:48 impressive:1 etc:1 align:3 add:1 recent:1 showed:1 irrelevant:1 certain:1 binary:2 meeting:1 accomplished:1 exploited:1 muller:1 morgan:1 greater:1 employed:1 r0:1 converge:1 determine:3 ii:2 branch:1 full:1 rj:1 segmented:2 characterized:4 retrieval:1 equally:1 histogram:2 normalization:1 represent:4 background:10 modality:2 publisher:1 sr:2 tend:6 effectiveness:1 ideal:5 split:1 variety:1 fit:2 zi:2 lesser:1 avenue:1 intensive:1 whether:2 useful:3 amount:1 transforms:1 dark:1 hue:1 ph:1 zabih:1 simplest:1 estimated:4 disjoint:1 blue:2 discrete:1 shall:1 promise:2 vol:1 four:1 pnew:2 pj:9 neither:1 registration:1 graph:1 sum:1 parameterized:1 almost:1 patch:3 followed:2 correspondence:2 oracle:1 annual:1 adapted:1 ri:2 scene:1 performing:1 alternate:1 belonging:1 across:5 leg:2 invariant:5 iccv:1 computationally:1 equation:3 previously:2 discus:1 unusual:2 gaussians:5 decomposing:3 experimentation:1 eight:3 hierarchical:3 rors:3 occurrence:5 matsakis:1 original:1 thomas:1 clustering:5 include:1 linguistics:1 restrictive:1 approximating:1 malik:1 added:1 quantity:1 diagonal:1 exhibit:1 september:1 simulated:1 chris:1 extent:1 collected:2 assuming:2 erik:1 modeled:2 useless:1 ratio:1 minimizing:1 difficult:1 unfortunately:1 negative:1 vertical:1 observation:2 enabling:1 viola:2 variability:1 head:5 rn:1 introduced:1 pair:5 required:1 kl:2 fort:1 product1:1 nip:1 address:1 able:1 indoor:1 including:1 memory:1 explanation:1 difficulty:1 rely:2 technology:1 brief:1 axis:1 created:1 extract:1 prior:3 relative:1 highlight:3 prototypical:1 proven:1 versus:1 thresholding:1 pi:18 pant:8 translation:1 row:2 english:1 bias:1 allow:1 institute:1 wide:1 template:31 face:1 subimage:1 tieu:2 depth:2 xn:2 valid:1 calculated:1 refinement:1 san:2 reconstructed:1 approximate:4 compact:1 implicitly:1 global:1 investigating:1 uai:1 francisco:1 latent:6 nature:1 robust:3 broadening:1 investigated:2 complex:1 noise:1 body:1 board:1 sub:3 position:1 pereira:1 candidate:1 weighting:1 specific:5 adding:1 effectively:2 gained:1 texture:4 conditioned:3 boston:1 simply:1 appearance:1 likely:3 explore:1 corresponds:1 determines:1 extracted:1 ma:2 conditional:4 towards:1 shared:1 considerable:2 determined:1 uniformly:1 invariance:1 meaningful:2 formally:1 support:1 dissimilar:1 absolutely:1 evaluate:1 handling:1 |
1,205 | 2,099 | A theory of neural integration in the
head-direction system
Richard H.R. Hahnloser , Xiaohui Xie and H. Sebastian Seung
Howard Hughes Medical Institute
Dept. of Brain and Cognitive Sciences
Massachusetts Institute of Technology
Cambridge, MA 02139
rhahnloser|xhxie|seung @mit.edu
Abstract
Integration in the head-direction system is a computation by which horizontal angular head velocity signals from the vestibular nuclei are integrated to yield a neural representation of head direction. In the thalamus, the postsubiculum and the mammillary nuclei, the head-direction
representation has the form of a place code: neurons have a preferred
head direction in which their firing is maximal [Blair and Sharp, 1995,
Blair et al., 1998, ?].
Integration is a difficult computation, given that head-velocities can vary
over a large range. Previous models of the head-direction system relied
on the assumption that the integration is achieved in a firing-rate-based
attractor network with a ring structure. In order to correctly integrate
head-velocity signals during high-speed head rotations, very fast synaptic
dynamics had to be assumed.
Here we address the question whether integration in the head-direction
system is possible with slow synapses, for example excitatory NMDA
and inhibitory GABA(B) type synapses. For neural networks with such
slow synapses, rate-based dynamics are a good approximation of spiking neurons [Ermentrout, 1994]. We find that correct integration during
high-speed head rotations imposes strong constraints on possible network architectures.
1 Introduction
Several network models have been designed to emulate the properties of head-direction
neurons (HDNs) [Zhang, 1996, Redish et al., 1996, Goodridge and Touretzky, 2000]. The
model by Zhang reproduces persistent activity during stationary head positions. Persistent
neural activity is generated in a ring-attractor network with symmetric excitatory and inhibitory synaptic connections. Independently, he and Redish et al. showed that integration
is possible by adding asymmetrical connections to the attractor network. They assumed
that the strength of these asymmetrical connections is modulated by head-velocity. When
the rat moves its head to the right, the asymmetrical connections induce a rightward shift
of the activity in the attractor network. A more plausible model without multiplicative
modulation of connections has been studied recently by Goodridge and Touretzky. There,
the head-velocity input has a modulatory influence on firing rates of intermittent neurons
rather than on connection strengths. The intermittent neurons are divided into two groups
that make spatially offset connections, one group to the right, the other to the left. The different types of neurons in the Goodridge and Touretzky model have firing properties that
are comparable to neurons in the various nuclei of the head-direction system.
What all these previous models have in common is that the integration is performed in an inherent double-ring network with very fast synapses (less than ms for
[Goodridge and Touretzky, 2000]). The connections made by one ring are responsible for
rightward turns and the connections made by the other ring are responsible for leftward
turns. In order to derive a network theory of integration valid for fast and slow synapses,
here we solve a simple double-ring network in the linear and in the saturated regimes.
An important property of the head-direction system is that the integration be linear over a
large range of head-velocities. We are interested in finding those type of synaptic connections that yield a large linear range and pose our findings as predictions on optimal network
architectures. Although our network is conceptually simpler than previous models, we
show that using two simple read-out methods, averaging and extracting the maximum, it is
possible to approximate head-velocity independent tuning curves as observed in the Postsubiculum (PoS) and anticipatory responses in the anterior dorsal thalamus (ADN).
2 Definition of the model
We assume that the number of neurons in the double-ring network is large and write its
dynamics as a continuous neural field
where
(1)
(2)
.-/
102.354
398:
'76
!#" $&%
(')
'+*
$),
('&
-/
;
102.3 4
398:=<
*
6
$ %
('&
! " $ ,
('&
'
0 @?BADC
E
> denotes a rectification nonlinearity.
and
are the firing rates of
"> :
neurons in the left and right ring, respectively. The quantities D and F represent
synaptic
). is a
and
activations (amount of neurotransmitter release3G4 caused3 by the
firing
rates
354H
3
'
6
6
synaptic
time
constant.
The
vestibular
inputs
and
are
purely
excitatory,
354JI
3IK354
3
'
6
. For simplicity, we assume that 6
is proportional to angular headvelocity. The synaptic connection profiles $ % between neurons on the same ring and $ ,
between neurons on different rings are given by:
LNM4OJM QPSRT
UWVB4LV XPGRT S<
(3)
$)%
$),
M4 M
VB4
V
*
, ,
and
define
the
intra
and
inter-ring
connection
strengths.
is
the
intra-ring
connection offset and the inter-ring offset.
3 Integration
3O
E . In
When the animal is not moving, the vestibular inputs to the two rings are equal, 6
this case, within a certain range of synaptic connections, steady bumps of activities appear
on the two rings. When
the head of the animal rotates, the activity bumps travel
3
3 at a velocity
determined by 6 . For perfect integration, should be proportional to 6
over the full
range of possible head-velocities. This is a difficult computational problem, in particular
for slow synapses.
4 Small head-velocity approximation
3
E ), the two stationary bumps of synaptic activation
When the head is not rotating ( 6
are of the form
0
L
0
4
4
'
'
and
(4)
" PGRT
(')
" PSR T
(')
:
:
4
where
is the current head direction and is the offset between the two bumps. How to
calculate , and is shown in the Appendix. The half width of these bumps is given by
WA PPSRT
S<
(5)
3 3S4
When the angular head velocity is small ( 6
), we linearize the dynamics around
the stationary solution Eq. (4), see Appendix. We find that
PGRT
0
4
G
('
('&
'
' '
(6)
" '
:
PGRT
;
10
4
F
('
('&
'
'J
(7)
"
:
M4
M
'
VB4F
'
V
PGRT *'
PGRT
T
where the velocity is given by
M XT
1*
3
"
6
and
'
M
'J
'
M4
T *
!
T
0
VB4 T
'
(8)
<
(9)
(10)
Equation (8) is the desired
result, relating the velocity of the two bumps to the differential
3
vestibular input 6 . In Fig. 1 we show simulation results using slow synapses (
E ms). The integration is V linear
Mover almost the entire range of head-velocities (up to
more than EE ! ) when
, i.e., when the amplitudes
V
NM of inter-ring and intra-ring
connections are equal. We point out that the condition
V cannot
4 directly be deduced
E ) was necessary to
from the above formulas, some empirical tuning (for
3 example
achieve this large range of linearity (large both in 6 and ).
"
# $%
'&%(
When the bumps move, their amplitudes tend to decrease. Fig. 1d shows the peak firing
rates of neurons in the two rings as a function of vestibular input. As can be seen, the firing
rates are a linear function of vestibular input, in agreement with equations 17 and 18 of the
Appendix. However, a linear firing-rate modulation by head velocity is not universal, for
some parameters we have seen asymmetrically head-velocity tuning, with a preference for
small head velocities (not shown).
a.
b.
800
600
800
Simulation
Theory
600
400
v (degrees/sec)
v (degrees/sec)
400
Simulation
Theory
200
0
?200
200
0
?200
?400
?400
?600
?600
?800
?1
?0.5
0
? b/b0
0.5
?800
?1
1
c.
?0.5
0.5
1
0.5
1
d.
75
600
Simulation
Theory
Left
Right
70
Firing rate (Hz)
400
v (degrees/sec)
0
? b/b0
200
0
?200
65
60
55
?400
?600
?1
?0.5
0
? b/b0
0.5
1
50
?1
?0.5
0
? b/b0
3 3G4
Figure 1: Velocity
V of
M activity
VB4 as a function of vestibular input 6 V . M a. Sublinear
bumps
E .V b. Supralinear
E ,
,*
,*
,
integration.
V 4
M
integration.
V 4
E . c. Linear (perfect) integration.
E . d. Head-velocity
,*
,
dependent modulation of firing rates
- (on the right and on the left ring). Same parameters
E ms. *
as in c.
, and
.
"
$
" $
" $
$
$
5 Saturating velocity
3
When 6 is very large, at some point, the left ring becomes inactive. Because inactivating
the left ring means that the push-pull competition between the two rings is minimized, we
are able to determine the saturating velocity of the double-ring network. The saturating
velocity is given by the on-ring connections $ % . Define
M4 @M XPGRT
('+*
$
M4 @M XPGRT
9A
* PSR T
* T
9A
* $ %
$&%
'
M4 M XPGRT
* PGRT
. Now, let
be the steady solution of a ring
where $)%
'
network with symmetric connections $&%
. By differentiating, it follows that
is the solution of a ring network with connections $
. Hence, the saturating
velocity
is given by
9A
* <
(11)
Notice that a traveling solution may not always exist if one ring is inactive (this is the case
when there are no intra-ring excitatory connections). However, even without a traveling
solution, equation (11) remains valid. In Figs. 1a and b, the saturating velocity is indicated
E !
by
dotted lines, in Fig. 1a we find
and in Fig. 1b
E the horizontal
.
" $%
'&%(
$'
'&'(
6 ADN and POs neurons
Goodridge and Touretzky?s integrator model was designed to emulate details of neuronal
tuning as observed in the different areas of the head-direction system. Wondering whether
the simple double ring studied here can also reproduce
multiple tuning curves, we analyze
simple read-out methods of the firing rates and . What we find is that two readout methods can indeed approximate response behavior resembling that of ADN and POs
neurons.
ADN
?BADC neurons:
By reading out firing rates using a maximum operation,
!
, anticipatory head-direction tuning arises due to the fact that there is
an activity offset between the two rings, equation (13). When the head turns to the right,
the activity on the right ring is larger than on the
left ring and so the tuning of
is
biased to the right. Similarly, for left turns,
is biased to the left. Thus, the activity
offset between the two rings leads to an anticipation time for ADN neurons, see Figure
2. Because, by assumption is head-velocity independent, it follows
that
is inversely
proportional to head-velocity (assuming perfect integration),
. In other words,
the anticipation time tends to be smaller for fast head rotations and larger for slow head
rotations.
D
POs
By reading out the double ring activity as an average,
neurons:
, neurons in POs do not have any anticipation time: because averaging is a symmetric
operation, all information about the direction of head rotations is lost.
Right turn
Left turn
Left ring
Right ring
Firing Rate
Max
Average
0
90
180
270
Head?direction (degs)
360
Figure 2: Snapshots of the activities on the two rings (top). Reading out the activities by
averaging and by a maximum operation (bottom).
7 Discussion
Here we discuss how the various connection parameters contribute to the double-ring network to function as an integrator. In particular we
3 discuss how parameters have to be tuned
in order to yield an integration that is large in 6 and in .
: By assumption the synaptic time constant is large. has the simplest effect
of all parameters on the integrator properties. According to equation (8), scales
the range of . Notice that if were small, a large range of could be trivially
achieved. The art here is to achieve this with large .
* : The connection offset * between neurons receiving similar vestibular input
is the sole parameter besides determing the saturating head-velocity, beyond
which integration is impossible. According to equation (11), the saturating velocity is large if * is close to E (we want the saturating velocity to be large). In
other words, for good integration, excitatory connections should be strongest (or
inhibitory connections weakest) for neuron pairs with preferred head-directions
differing by close to E .
: The connection offset between neurons receiving different
- vestibular input
determines the anticipation time of thalamic neurons. If
is large, then ,
$
$
the activity offset in equation (13) is large. And,- because is proportional to
(assuming perfect integration), we conclude that should preferentially be large
(close to E - ) if is to be large. Notice that by equation (8), the range of is not
affected by .
VB4
V
and VB:4 The inter-ring connections should
V=4 be mainly excitatory, which imE was found to be optimal). The
plies that
should not be too negative (
3
intuitive reason is the following. We want the integration to be as linear in 6 as
possible, which means that we want our linear expansions (6) and (7) to deviate
as little as possible from (4). Hence, the differential gain between the two rings
should be small, which is the case when the two rings excite
each other. The inter3
ring excitation makes sure, even for large values of 6 , that there are comparable
activity levels on the two rings. This is one of the main points of this study.
M4
M
and
: The intra-ring connections should be mainly inhibitory, which implies
M4
that
should be strongly negative. The reason for this is that inhibition is necessary to result in proper and stable integration. Since
M 4 inhibition cannot come
from the inter-ring connections,
it
has
to
come
from
M
V . Notice also that according to equation (15),
cannot be much larger than
. If this were the case,
the persistent activity in the no head-movement caseV would
M become unstable. For
linear integration we have found that the condition
is necessary; small
deviations from this condition cause the integrator to become sub- or supralinear.
$
8 Conclusion
We have presented a theory for integration in the head-direction system with slow synapses.
We have found that in order to achieve a large range of linear integration, there should be
strong excitatory connections between neurons with dissimilar head-velocity tuning and
inhibitory connections between neurons with similar head-velocity tuning (see the discussion). Similar to models of the occulomotor integrator [Seung, 1996], we have found that
linear
can only be achieved by precise tuning of synaptic weights (for example
V
Nintegration
M
).
Appendix
, it is convenient to go into a moving
To study the traveling pulse solution with velocity
B'
coordinate frame by the change of variables
. The stationary solution in the
moving frame reads
W
'
and '
(12)
E . In order to find the fixed points of equation
(12),
we use the ansatz (4) and
Set
equate the coefficients of the 3 Fourier modes T , PGRT
and the -independent mode.
This leads to
A PT
M
V
QT 1*
'
(13)
3 4
(14)
M4L.VB4F 4
'
' PGRT
M XPGRT
V
M T
8
*
'
1*
(15)
where the functions
4
U
D4
and
D " T
are given by
'&
PGRT
10
D "
'
T
0<
E . Eq. (13) determines the
The above set of equations fully characterize the solution for
offset between the two rings, eq. (15) determines the threshold
, eq. (14) the amplitude
and eq. (5) the bias .
3
When the vestibular input 6
is small, we assume that the perturbed solution around
and takes the form:
4
PGRT
(')
'
PGRT
4F
S<
')
'
We linearize the dynamics (12) (to first order in
) and equate the Fourier coefficients.
This leads to
M T
T
0
*
T
'
(16)
"
!
' and ' . We determine and by solving the
where
'
PSR T
'
4F ' .
'
linearized dynamics of the differential mode
Comparing once more the Fourier coefficients leads to
6 3 " 3
' ' T
0 0
(17)
6 "
' ' T
(18)
M 4
V 4
'
. By substituting and into Eq. (16), we find equation (8).
where
References
[Blair et al., 1998] Blair, H., Cho, J., and Sharp, P. (1998). Role of the lateral mammillary
nucleus in the rat head direction circuit: A combined single unit recording and lesion
study. Neuron, 21:1387?1397.
[Blair and Sharp, 1995] Blair, H. and Sharp, P. (1995). Anticipatory head diirection signals
in anterior thalamus: evidence for a thalamocortical circuit that integrates angular head
motion to compute head direction. The Journal of Neuroscience, 15(9):6260?6270.
[Ermentrout, 1994] Ermentrout, B. (1994). Reduction of conductance-based models with
slow synapses to neural nets. Neural Computation, 6:679?695.
[Goodridge and Touretzky, 2000] Goodridge, J. and Touretzky, D. (2000). Modeling attractor deformation in the rodent head-direction system. The Journal of Neurophysiology, 83:3402?3410.
[Redish et al., 1996] Redish, A., Elga, A. N., and Touretzky, D. (1996). A coupled attractor
model of the rodent head direction system. Network: Computation in Neural Systems,
7:671?685.
[Seung, 1996] Seung, H. S. (1996). How the brain keeps the eyes still. Proc. Natl. Acad.
Sci. USA, 93:13339?13344.
[Zhang, 1996] Zhang, K. (1996). Representation of spatial orientation by the intrinsic
dynamics of the head-direction cell ensemble: A theory. J. Neurosci., 16(6):2112?2126.
| 2099 |@word neurophysiology:1 pulse:1 linearized:1 simulation:4 reduction:1 tuned:1 current:1 comparing:1 anterior:2 activation:2 designed:2 stationary:4 half:1 contribute:1 preference:1 simpler:1 zhang:4 differential:3 become:2 persistent:3 g4:1 inter:5 indeed:1 behavior:1 brain:2 integrator:5 little:1 jm:1 becomes:1 linearity:1 circuit:2 what:2 differing:1 finding:2 unit:1 medical:1 appear:1 inactivating:1 tends:1 acad:1 wondering:1 firing:14 modulation:3 studied:2 goodridge:7 range:11 responsible:2 hughes:1 lost:1 area:1 universal:1 empirical:1 convenient:1 word:2 induce:1 elga:1 anticipation:4 cannot:3 close:3 influence:1 impossible:1 xiaohui:1 resembling:1 go:1 independently:1 simplicity:1 pull:1 coordinate:1 agreement:1 velocity:32 observed:2 bottom:1 role:1 calculate:1 readout:1 decrease:1 movement:1 seung:5 ermentrout:3 dynamic:7 solving:1 purely:1 rightward:2 po:5 emulate:2 various:2 neurotransmitter:1 fast:4 larger:3 plausible:1 solve:1 net:1 maximal:1 achieve:3 intuitive:1 competition:1 double:7 perfect:4 ring:46 derive:1 linearize:2 pose:1 qt:1 sole:1 b0:4 eq:6 strong:2 implies:1 blair:6 come:2 direction:22 correct:1 around:2 bump:8 substituting:1 vary:1 proc:1 travel:1 integrates:1 mit:1 always:1 rather:1 mainly:2 dependent:1 integrated:1 entire:1 reproduce:1 interested:1 orientation:1 animal:2 art:1 integration:26 spatial:1 field:1 equal:2 once:1 minimized:1 adn:5 richard:1 inherent:1 ime:1 mover:1 attractor:6 conductance:1 intra:5 saturated:1 natl:1 necessary:3 rotating:1 desired:1 xhxie:1 deformation:1 modeling:1 deviation:1 too:1 characterize:1 perturbed:1 cho:1 combined:1 deduced:1 peak:1 receiving:2 ansatz:1 nm:2 cognitive:1 redish:4 sec:3 coefficient:3 multiplicative:1 performed:1 analyze:1 inter3:1 relied:1 thalamic:1 equate:2 ensemble:1 yield:3 conceptually:1 synapsis:9 strongest:1 touretzky:8 sebastian:1 synaptic:10 definition:1 gain:1 massachusetts:1 nmda:1 amplitude:3 xie:1 response:2 anticipatory:3 strongly:1 angular:4 traveling:3 horizontal:2 mode:3 indicated:1 usa:1 effect:1 asymmetrical:3 hence:2 spatially:1 symmetric:3 read:3 during:3 width:1 steady:2 excitation:1 rat:2 m:3 d4:1 motion:1 mammillary:2 recently:1 common:1 rotation:5 spiking:1 ji:1 he:1 relating:1 cambridge:1 tuning:10 postsubiculum:2 trivially:1 similarly:1 nonlinearity:1 had:1 moving:3 stable:1 inhibition:2 showed:1 leftward:1 certain:1 seen:2 determine:2 signal:3 full:1 multiple:1 thalamus:3 divided:1 qpsr:1 prediction:1 represent:1 achieved:3 cell:1 want:3 biased:2 sure:1 hz:1 tend:1 recording:1 extracting:1 architecture:2 shift:1 inactive:2 whether:2 cause:1 modulatory:1 amount:1 s4:1 simplest:1 exist:1 inhibitory:5 notice:4 dotted:1 neuroscience:1 correctly:1 write:1 affected:1 group:2 threshold:1 place:1 almost:1 appendix:4 vb:1 comparable:2 activity:15 strength:3 constraint:1 fourier:3 speed:2 according:3 gaba:1 smaller:1 rectification:1 equation:12 remains:1 turn:6 discus:2 operation:3 denotes:1 top:1 move:2 question:1 quantity:1 rotates:1 lateral:1 sci:1 unstable:1 reason:2 assuming:2 code:1 besides:1 preferentially:1 difficult:2 negative:2 proper:1 neuron:25 snapshot:1 howard:1 precise:1 head:54 frame:2 intermittent:2 sharp:4 pair:1 connection:29 vestibular:10 address:1 able:1 beyond:1 regime:1 reading:3 max:1 technology:1 inversely:1 eye:1 coupled:1 deviate:1 fully:1 sublinear:1 proportional:4 nucleus:4 integrate:1 degree:3 imposes:1 excitatory:7 thalamocortical:1 bias:1 institute:2 differentiating:1 curve:2 valid:2 made:2 approximate:2 supralinear:2 preferred:2 keep:1 reproduces:1 assumed:2 conclude:1 excite:1 continuous:1 expansion:1 main:1 neurosci:1 profile:1 lesion:1 determing:1 neuronal:1 fig:5 slow:8 sub:1 position:1 ply:1 formula:1 xt:1 badc:2 offset:10 weakest:1 evidence:1 intrinsic:1 adding:1 push:1 rodent:2 psr:4 saturating:8 determines:3 ma:1 hahnloser:1 change:1 determined:1 averaging:3 asymmetrically:1 modulated:1 dorsal:1 arises:1 dissimilar:1 dept:1 |
1,206 | 21 | 674
PA'ITERN CLASS DEGENERACY IN AN UNRESTRICfED STORAGE
DENSITY MEMORY
Christopher L. Scofield, Douglas L. Reilly,
Charles Elbaum, Leon N. Cooper
Nestor, Inc., 1 Richmond Square, Providence, Rhode Island,
02906.
ABSTRACT
The study of distributed memory systems has produced a
number of models which work well in limited domains.
However, until recently, the application of such systems to realworld problems has been difficult because of storage limitations,
and their inherent architectural (and for serial simulation,
computational) complexity.
Recent development of memories
with unrestricted storage capacity and economical feedforward
architectures has opened the way to the application of such
systems to complex pattern recognition problems.
However,
such problems are sometimes underspecified by the features
which describe the environment, and thus a significant portion
of the pattern environment is often non-separable.
We will
review current work on high density memory systems and their
network implementations.
We will discuss a general learning
algorithm for such high density memories and review its
application to separable point sets. Finally, we will introduce an
extension of this method for learning the probability
distributions of non-separable point sets.
INTRODUcnON
Information storage in distributed content addressable
memories has long been the topic of intense study.
Early
research focused on the development of correlation matrix
memories 1, 2, 3, 4. Workers in the field found that memories of
this sort allowed storage of a number of distinct memories no
larger than the number of dimensions of the input space.
Further storage beyond this number caused the system to give
an incorrect output for a memorized input.
@ American Institute of Physics 1988
675
Recent work on distributed memory systems has focused on
single layer, recurrent networks.
Hopfield 5, 6 introduced a
method for the analysis of settling of activity in recurrent
networks.
This method defined the network as a dynamical
system for which a global function called the 'energy' (actually a
Liapunov function for the autonomous system describing the
Hopfield showed that
state of the network) could be defined.
flow in state space is always toward the fixed points of the
dynamical system if the matrix of recurrent connections satisfies
certain conditions.
With this property, Hopfield was able to
define the fixed points as the sites of memories of network
acti vity.
Like its forerunners, the Hopfield network is limited in
storage capacity. Empirical study of the system found that for
randomly chosen memories, storage capacity was limited to m ~
O.lSN, where m is the number of memories that could be
accurately recalled, and N is the dimensionality of the network
(this has since been improved to m ~ N, 7, 8). The degradation of
memory recall with increased storage density is directly related
to the proliferation in the state space of unwanted local minima
which serve as basins of flow.
UNRESTRICIEn STORAGE DENSITY MEMORIES
Bachman et al. 9 have studied another relaxation system
similar in some respects to the Hopfield network. However, in
contrast to Hopfield, they have focused on defining a dynamical
system in which the locations of the minima are explicitly
known.
In particular, they have chosen a system with a Liapunov
function given by
E = -IlL ~
Qj I Il- Xj I - L,
(1)
J
where E is the total 'energy' of the
describing the initial network activity
and Xj' the site of the jth memory, for
parameter related to the network size.
= Xj for some memory j according to
network, Il (0) is a vector
caused by a test pattern,
m memories in RN. L is a
Then 1l(0) relaxes to Il(T)
676
(2)
This system is isomorphic to the classical electrostatic potential
between a positive (unit) test charge, and negative charges Qj at
the sites Xj (for a 3-dimensional input space, and L = 1). The Ndimensional Coulomb energy function then defines exactly m
basins of attraction to the fixed points located at the charge sites
Xj.
It can been shown that convergence to the closest distinct
memory is guaranteed, independent of the number of stored
memories m, for proper choice of Nand L 9, to.
Equation 1 shows that each cell receives feedback from the
network in the form of a scalar
~ Q-I Jl- x-I- L
J J
J
?
(3)
Importantly, this quantity is the same for all cells; it is as if a
single virtual cell was computing the distance in activity space
between the current state and stored states. The result of the
computation is then broadcast to all of the cells in the network.
A 2-layer feedforward network implementing such a system has
been described elsewhere 10 .
The connectivity for this architecture is of order m?N, where
m is the number of stored memories and N is the dimensionality
of layer 1.
This is significant since the addition of a new
memory m' = m + 1 will change the connectivity by the addition
of N + 1 connections, whereas in the Hopfield network, addition
of a new memory requires the addition of 2N + 1 connections.
An equilibrium feedforward network with similar properties
has been under investigation for some time 11. This model does
not employ a relaxation procedure, and thus was not originally
framed in the language of Liapunov functions.
However, it is
possible to define a similar system if we identify the locations of
the 'prototypes' of this model as? the locations in state space of
potentials which satisfy the following conditions
Ej
= -Qj lRo for I j.t - Xj I < Aj
=0
for I fl - Xj I > A].
(4)
677
where Ro is a constant.
This form of potential is often referred to as the 'square-well'
potential. This potential may be viewed as a limit of the Ndimensional Coulomb potential, in which the l/R (L = l) well is
replaced with a square well (for which L ? l). Equation 4
describes an energy landscape which consists of plateaus of zero
potential outside of wells with flat, zero slope basins. Since the
landscape has only flat regions separated by discontinuous
boundaries, the state of the network is always at equilibrium,
and relaxation does not occur. For this reason, this system has
been called an equilibrium model. This model, also referred to
as the Restricted Coulomb Energy (RCE)14 model, shares the
property of unrestricted storage density.
LEARNING IN HIGH DENSITY MEMORIES
A simple learning algorithm for the placement of the wells has
been described in detail elsewhere 11, 12.
Figurel: 3-layer feedforward network. Cell i
computes the quantity IJl - xii and compares
to internal threshold Ai.
678
Reilly et. al. have employed a three layer feedforward
network (figure 1) which allows the generalization of a content
addressable memory to a pattern classification memory.
Because the locations of the minima are explicitly known in the
equilibrium model, it is possible to dynamically program the
energy function for an arbitrary energy landscape. This allows
the construction of geographies of basins associated with the
classes constituting the pattern environment. Rapid learning of
complex, non-linear, disjoint, class regions is possible by this
method 12, 13.
LEARNING NON-SEPARABLE CLASS REGIONS
Previous studies have focused on the acquisition of the
geography and boundaries of non-linearly separable point sets.
However, a method by which such high density models can
acquire the probability distributions of non-separable sets has
not been described.
Non-separable sets are defined as point sets in the state
space of a system which are labelled with multiple class
affiliations.
This can occur because the input space has not
carried all of the features in the pattern environment, or because
the pattern set itself is not separable. Points may be degenerate
with respect to the explicit features of the space, however they
may have different probability distributions within the
environment.
This structure in the environment is important
information for the identification of patterns by such memories
10 the presence of feature space degeneracies.
We now describe one possible mechanism for the acquisition
of the probability distribution of non-separable points.
It is
assumed that all points in some region R of the state space of the
network are the site of events Jl (0, Ci ) which are examples of
pattern classes C = {C 1 , ... , CM }. A basin of attraction, xk( C i ),
defined by equation 4, is placed at each site fl(O, Ci ) unless
(5)
that is, unless a memory at Xj (of the class Ci ) already contains
fl(O, Ci )? The initial values of Qo and Ro at xk(Ci) are a constant for
all sites Xj. Thus as events of the classes C 1 , ... , C M occur at a
particular site in R, multiple wells are placed at this location.
679
If a well x/ C i) correctly covers an event Jl (0, C i ), then the
charge at that site (which defines the depth of the well) is
incremented by a constant amount ~ Q o. In this manner, the
region R is covered with wells of all classes {C 1 , ... , C M }, with the
depth of well XiCi) proportional to the frequency of occurence of
C i at Xj.
The architecture of this network is exactly the same as that
already described. As before, this network acquires a new cell
for each well placed in the energy landscape. Thus we are able
to describe the meaning of wells that overlap as the competition
by multiple cells in layer 2 in firing for the pattern of activity in
the input layer.
APPLICATIONS
This system has been applied to a problem in the area of risk
assessment in mortgage lending. The input space consisted of
feature detectors with continuous firing rates proportional to the
values of 23 variables in the application for a mortgage. For this
set of features, a significant portion of the space was nonseparable.
Figures 2a and 2b illustrate the probability distributions of
high and low risk applications for two of the features. It is clear
that in this 2-dimensional subspace, the regions of high and low
risk are non-separable but have different distributions.
t-----------#llir----- Prob. = 1.0.
1000 Patterns
Prob.
0.0
Feature 1
= 0.5
1.0
Figure 2a: Probability distribution for High
and Low risk patterns for feature 1.
680
=
Prob.
1.0.
t 000 Patterns
1-----1----\---------
Prob.
0.0
= 0.5
1.0
Feature 2
Figure 2b: Probability distribution for High
and Low risk patterns for feature 2.
Figure 3 depicts the probability distributions acquired by
the system for this 2-dimensional subspace.
In this image,
circle radius is proportional to the degree of risk: Small circles
are regions of low risk, and large circles are regions of high
risk.
00
o
o
V
0 0
0
0
o
0:>0
t?.
0
0
00
0
o
o
00
0
0
0
00
0
o
Feature 1
Figure 3: Probability distribition for Low and
High risk.
Small circles indicate low risk
regIons and large circles indicate high risk
regions.
681
Of particular interest is the clear clustering of high and low risk
regions in the 2-d map. Note that the regions are in fact nonlinearly separable.
DISCUSSION
We have presented a simple method for the acquisition of
probability distributions in non-separable point sets.
This
method generates an energy landscape of potential wells with
depths that are proportional to the local probability density of
the classes of patterns in the environment. These well depths
set the probability of firing of class cells In a 3-layer
feedforward network.
Application of this method to a problem in risk assessment
has shown that even completely non-separable subspaces may
be modeled with surprising accuracy.
This method improves
pattern classification in such problems with little additional
computational burden.
This algorithm has been run in conjunction with the method
described by Reilly et. al. II for separable regions. This combined
system is able to generate non-linear decision surfaces between
the separable zones, and approximate the probability
distributions of the non-separable zones in a seemless manner.
Further discussion of this system will appear in future reports.
Current work is focused on the development of a more
general method for modelling the scale of variations in the
distributions.
Sensitivity to this scale suggests that the
transition from separable to non-separable regions is smooth
and should not be handled with a 'hard' threshold.
ACKNOWLEDGEMENTS
We would like to thank Ed Collins and Sushmito Ghosh for their
significant contributions to this work through the development
of the mortgage risk assessment application.
REFERENCES
[1] Anderson, J .A.: A simple neural network generating an
interactive memory. Math. Biosci. 14, 197-220 (1972).
682
[2] Cooper, L.N.: A possible organization of animal memory and
learning. In: Proceedings of the Nobel Symposium on Collective
Properties of Physical Systems, Lundquist, B., Lundquist, S.
(eds.). (24), 252-264 London, New York: Academic Press 1973.
[3] Kohonen, T.: Correlation matrix memories.
IEEE Trans.
Comput. 21, 353-359 (1972).
[4] Kohonen, T.: Associative memory - a system-theoretical
approach. Berlin, Heidelberg, New York: Springer 1977.
[5] Hopfield, J.J.: Neural networks and physical systems with
emergent collective computational abilities. Proc. Natl. Acad. Sci.
USA 79, 2554-2558 (April 1982).
[6] Hopfield, J.J.: Neurons with graded response have collective
computational properties like those of two-state neurons. Proc.
Natl. Acad. Sci. USA 81, 2088-3092 (May, 1984).
[7] Hopfield, J.J., Feinstein, D.I., Palmer, R.G.: 'Unlearning' has a
stabilizing effect in collective memories. Nature 304, 158-159
(July 1983).
[8] Potter, T.W.: Ph.D. Dissertation in advanced technology,
S.U.N.Y. Binghampton, (unpublished).
[9] Bachmann, C.M., Cooper, L.N., Dembo, A., Zeitouni, 0.: A
relaxation model for memory with high density storage. to be
published in Proc. Nati. Acad. Sci. USA.
[10] Dembo, A., Zeitouni, 0.: ARO Technical Report, Brown
University, Center for Neural Science, Pr0vidence, R.I., (1987),
also submitted to Phys. Rev. A.
[11] Reilly, D.L., Cooper, L.N., Elbaum, C.: A neural model for
category learning. BioI. Cybern. 45, 35 -41 (1982).
[12] Reilly, D.L., Scofield, C., Elbaum, C., Cooper, L.N.: Learning
system architectures composed of multiple learning modules. to
appear in Proc. First In1'1. Conf. on Neural Networks (1987).
[13] Rimey, R., Gouin, P., Scofield, C., Reilly, D.L.: Real-time 3-D
object classification using a learning system. Intelligent Robots
and Computer Vision, Proc. SPIE 726 (1986).
[14] Reilly, D.L., Scofield, C. L., Elbaum, C., Cooper, L.N: Neural
Networks with low connectivity and unrestricted memory
storage density. To be published.
| 21 |@word simulation:1 bachman:1 initial:2 contains:1 current:3 surprising:1 liapunov:3 xk:2 dembo:2 dissertation:1 math:1 lending:1 location:5 symposium:1 incorrect:1 consists:1 acti:1 unlearning:1 introduce:1 manner:2 acquired:1 rapid:1 proliferation:1 nonseparable:1 little:1 cm:1 elbaum:4 ghosh:1 charge:4 interactive:1 unwanted:1 exactly:2 ro:2 unit:1 appear:2 positive:1 before:1 local:2 limit:1 acad:3 firing:3 rhode:1 mortgage:3 studied:1 dynamically:1 suggests:1 limited:3 palmer:1 procedure:1 addressable:2 area:1 empirical:1 reilly:7 storage:13 risk:14 cybern:1 map:1 center:1 focused:5 stabilizing:1 rimey:1 attraction:2 importantly:1 autonomous:1 variation:1 construction:1 pa:1 recognition:1 located:1 underspecified:1 module:1 region:14 incremented:1 environment:7 complexity:1 serve:1 completely:1 hopfield:10 emergent:1 separated:1 distinct:2 describe:3 london:1 outside:1 larger:1 ability:1 itself:1 associative:1 aro:1 kohonen:2 degenerate:1 competition:1 convergence:1 generating:1 object:1 illustrate:1 recurrent:3 indicate:2 radius:1 discontinuous:1 opened:1 memorized:1 virtual:1 implementing:1 generalization:1 geography:2 investigation:1 extension:1 equilibrium:4 early:1 proc:5 always:2 ej:1 conjunction:1 modelling:1 richmond:1 contrast:1 nand:1 classification:3 ill:1 development:4 animal:1 field:1 future:1 report:2 intelligent:1 inherent:1 employ:1 randomly:1 composed:1 nestor:1 replaced:1 organization:1 interest:1 natl:2 worker:1 intense:1 unless:2 circle:5 theoretical:1 increased:1 cover:1 stored:3 providence:1 combined:1 density:11 sensitivity:1 physic:1 connectivity:3 broadcast:1 conf:1 american:1 potential:8 inc:1 satisfy:1 caused:2 explicitly:2 portion:2 sort:1 slope:1 contribution:1 square:3 il:3 accuracy:1 identify:1 landscape:5 identification:1 accurately:1 produced:1 economical:1 published:2 submitted:1 plateau:1 detector:1 phys:1 ed:2 energy:9 acquisition:3 frequency:1 associated:1 spie:1 degeneracy:2 recall:1 dimensionality:2 improves:1 actually:1 originally:1 response:1 improved:1 april:1 anderson:1 until:1 correlation:2 receives:1 christopher:1 qo:1 assessment:3 defines:2 aj:1 usa:3 effect:1 consisted:1 brown:1 acquires:1 ijl:1 meaning:1 image:1 recently:1 charles:1 physical:2 jl:3 significant:4 biosci:1 ai:1 framed:1 language:1 robot:1 surface:1 electrostatic:1 closest:1 recent:2 showed:1 introducnon:1 certain:1 vity:1 affiliation:1 minimum:3 unrestricted:3 additional:1 employed:1 july:1 ii:1 multiple:4 smooth:1 technical:1 academic:1 long:1 serial:1 vision:1 sometimes:1 cell:8 addition:4 whereas:1 flow:2 presence:1 feedforward:6 relaxes:1 xj:10 architecture:4 prototype:1 qj:3 handled:1 lsn:1 york:2 covered:1 clear:2 amount:1 ph:1 category:1 generate:1 lro:1 disjoint:1 correctly:1 xii:1 threshold:2 douglas:1 relaxation:4 realworld:1 prob:4 run:1 architectural:1 decision:1 layer:8 fl:3 guaranteed:1 activity:4 occur:3 placement:1 flat:2 generates:1 leon:1 separable:18 according:1 describes:1 island:1 rev:1 restricted:1 equation:3 discus:1 describing:2 mechanism:1 feinstein:1 coulomb:3 clustering:1 rce:1 zeitouni:2 graded:1 classical:1 already:2 quantity:2 in1:1 subspace:3 distance:1 thank:1 berlin:1 capacity:3 sci:3 topic:1 toward:1 reason:1 nobel:1 potter:1 modeled:1 acquire:1 difficult:1 negative:1 implementation:1 proper:1 collective:4 neuron:2 defining:1 rn:1 arbitrary:1 introduced:1 nonlinearly:1 unpublished:1 connection:3 recalled:1 trans:1 beyond:1 able:3 dynamical:3 pattern:16 program:1 memory:35 event:3 overlap:1 settling:1 ndimensional:2 advanced:1 technology:1 carried:1 occurence:1 review:2 acknowledgement:1 nati:1 limitation:1 proportional:4 degree:1 basin:5 share:1 elsewhere:2 placed:3 jth:1 scofield:4 institute:1 distributed:3 feedback:1 dimension:1 boundary:2 depth:4 transition:1 computes:1 constituting:1 approximate:1 global:1 assumed:1 continuous:1 nature:1 heidelberg:1 complex:2 domain:1 linearly:1 allowed:1 site:9 referred:2 depicts:1 cooper:6 explicit:1 comput:1 bachmann:1 burden:1 ci:5 scalar:1 springer:1 satisfies:1 bioi:1 viewed:1 labelled:1 content:2 change:1 hard:1 degradation:1 called:2 total:1 isomorphic:1 zone:2 internal:1 collins:1 |
1,207 | 210 | 100
Servan-Schreiber, Printz and Cohen
The Effect of Catecholamines on Performance:
From Unit to System Behavior
David Servan-Schreiber, Harry Printz and Jonathan D. Cohen
School of Computer Science and Department of Psychology
Carnegie Mellon University
Pittsburgh. PA 15213
ABSTRACT
At the level of individual neurons. catecholamine release increases the
responsivity of cells to excitatory and inhibitory inputs. We present a
model of catecholamine effects in a network of neural-like elements.
We argue that changes in the responsivity of individual elements do
not affect their ability to detect a signal and ignore noise. However.
the same changes in cell responsivity in a network of such elements do
improve the signal detection performance of the network as a whole. We
show how this result can be used in a computer simulation of behavior
to account for the effect of eNS stimulants on the signal detection
performance of human subjects.
1 Introduction
The catecholamines-norepinephrine and dopamine-are neuroactive substances that are
presumed to modulate information processing in the brain, rather than to convey discrete
sensory or motor signals. Release of norepinephrine and dopamine occurs over wide
areas of the central nervous system. and their post-synaptic effects are long lasting.
These effects consist primarily of an enhancement of the response of target cells to other
afferent inputs, inhibitory as well as excitatory (see [4] for a review).
Increases or decreases in catecholaminergic tone have many behavioral consequences
including effects on motivated behaviors. attention, learning and memory. and motor
The Effect of Catecholamines on Performance: From Unit to System Behavior
behavior. At the information processing level, catecholamines appear to affect the ability
to detect a signal when it is imbedded in noise (see review in [3]).
In terms of signal detection theory, this is described as a change in the performance of the
system. However, there is no adequate account of how these changes at the system level
relate to the effect of catecholamines on individual cells. Several investigators [5,12,2]
have suggested that catecholamine-mediated increases in a cell's responsivity can be
interpreted as a change in the cell's signal-to-noise ratio. By analogy, they proposed that
this change at the unit level may account for changes in signal detection performance at
the behavioral level.
In the first part of this paper we analyze the relation between unit responsivity, signal-tonoise ratio and signal detection performance in a network of neural elements. We start
by showing that the changes in unit responsivity induced by catecholamines do not result
in changes in signal detection performance of a single unit. We then explain how, in
spite of this fact, the aggregrate effect of such changes in a chain of units can lead to
improvements in the signal detection performance of the entire network.
In the second part, we show how changes in gain - which lead to an increase in the
signal detection performance of the network - can account for a behavioral phenomenon.
We describe a computer simulation of a network performing a signal detection task that
has been applied extensively to behavioral research: the continuous performance test.
In this simulation, increasing the responsivity of individual units leads to improvements
in performance that closely approximate the improvement observed in human subjects
under conditions of increased catecholaminergic tone.
2 Effect of Gain on a Single Element
We assume that the response of a typical neuron can be described by a strictly increasing
function !G(x) from real-valued inputs to the interval (0, 1). This function relates the
strength of a neuron's net afferent input x to its probability of firing, or activation. We
do not require that!G is either continuous or differentiable.
For instance, the family of logistics, given by
1
!G(x)
= 1 + e-(G%+B)
has been proposed as a model of neural activation functions [7,1]. These functions are
all strictly increasing, for each value of the gain G> 0, and all values of the bias B.
The potentiating effect of catecholamines on responsivity can be modelled as a change
in the shape of its activation function. In the case of the logistic, this is achieved by
increasing the value of G, as illustrated in Figure 1. However, our analysis applies to
any suitable family of functions, {fG}. We require only that each member function!G
is strictly increasing, and that as G -;. 00, the family {fG} converges monotonically to
101
102
Servan-Schreiber, Printz and Cohen
0.0
b==::::L::==-._:::::::::=-----I'---__
-6.0
....L...-_ _ _ _ _ _- - - '
-<lJ)
-2.0
OJ)
2J)
-IJ)
6J)
" (Nell""..,)
Figure 1: Logistic Activation Function, Used to Model the Response Function of Neurons. Positive
net inputs correspond to excitatory stimuli, negative net inputs correspond to inhibitory stimuli.
For the graphs drawn here, we set the bias B to -1. The asymmetry arising from a negative bias
is often found in the response function of actual neurons [6].
the unit step function Uo almost everywhere. 1 Here. Uo is defined as
0 for x < 0
uo(x) -- { 1 for x -> 0
This means that as G increases. the value !G(x) gets steadily closer to 1 if x > O. and
steadily closer to 0 if x < O.
2.1
Gain Does Not Affect Signal Detection Performance
Consider the signal detection performance of a network in which the response of a single
unit is compared with a threshold to determine the presence or absence of a signal. We
assume that in the presence of the signal. this unit receives a positive (excitatory) net
afferent input Xs. and in the absence of the signal it receives a null or negative (inhibitory)
input XA. When zero-mean noise is added to this quantity. in the presence as well as the
absent:e of the signal, the unit's net input in each case is distributed around Xs or XA
respectively. Therefore its response is distributed around !G(xs) or !G(XA) respectively
(see Figure 2).
In other words, the input in the case where the signal is present is a random variable
Xs ? with probability density function (pdt) PXs and mean Xs, and in the absence of the
signal it is the random variable XA? with pdf PXA and mean XA. These then determine
the random variables YGS =!G(Xs) and YGA =!G(XA). with pdfs PYas and PYGA' which
represent the response in the presence or absence of the signal for a given value of the
gain. Figure 2 shows examples of PYas and PYGA for two different values of G. in the case
where!G is the biased logistic.
If the input pdfs PXs and PXA overlap. the output pdfs PYas and PYGA will also overlap.
Thus for any given threshold () on the y-axis used to categorize the output as "signal
present" or "signal absent," there will be some misses and some false alarms. The best
1 A sequence of functions {gil} converges almost everywhere
diverges, or converges to the wrong value, is of measure zero.
to
the function g if the set of points where it
The Effect of Catecholamines on Performance: From Unit to System Behavior
01)
?21)
41)
21)
%
01)
?21)
21)
41)
61)
(Nelb'plll)
61)
% (Nelillplll)
---------p-~----
----~p~----------
Figure 2: Input and Output Probability Density Functions. The curves at the bottom are the pdfs
of the net input in the signal absent (left) and signal present (right) cases. The curves along the
y-axis are the response pdfs for each case; they are functions of the activation y, and represent the
distribution of outputs. The top graph shows the logistic and response pdfs for G = 0.5, B = -1;
the bottom graph shows them for G = 1. 0, B -1.
=
the system can do is to select a threshold that optimizes performance. More precisely,
the expected payoff or performance of the unit is given by
E(O) = A + a:.
Pr(YGS ~
0) - (3. Pr(YGA
~
0)
where A, a:, and (3 are constants that together reflect the prior probability of the signal,
and the payoffs associated with correct detections or hits, correct ignores, false alarms
and misses. Note that Pr(YGS ~ 0) and Pr(YGA ~ 0) are the probabilities of a hit and a
false alarm, respectively.
By solving the equation dE/dB = 0 we can determine the value 0* that maximizes E. We
call 0* the optimal threshold. Our first result is that for any activation function f that
satisfies our assumptions, and any fixed input pdfs PXs and PXA the unit's performance at
optimal threshold is the same. We call this the Constant Optimal Performance Theorem,
which is stated and proved in [10]. In particular, for the logistic, increasing the gain
G does not induce better performance. It may change the value of the threshold that
yields optimal performance, but it does not change the actual performance at optimum.
This is because a strictly increasing activation function produces a point-to-point mapping
between the distributions of input and output values. Since the amount of overlap between
103
104
Servan-Schreiber, Printz and Cohen
the two input pdfs PXs and PXA does not change as the gain varies, the amount of overlap
in the response pdfs does not change either, even though the shape of the response pdfs
does change when gain increases (see Figure 2). 2
3
Effect of Gain on a Chain of Elements
Although increasing the gain does not affect the signal detection performance of a single
element, it does improve the perfonnance of a chain of such elements. By a chain, we
mean an arrangement in which the output of the firs t unit provides the input to another
unit (see Figure 3). Let us call this second element the response unit We monitor the
output of this second unit to detennine the presence or absence of a signal.
Input Unit
x
Response Unit
z
y
v
Figure 3: A Chain of Units. The output of the unit receiving the signal is combined with noise
to provide input to a second unit, called the response unit. The activation of the response unit is
compared to a threshold to determine the presence or absence of the signal.
As in the previous discussion. noise is added to the net input to each unit in the chain
in the presence as well as in the absence of a signal. We represent noise as a random
variable V. with pdf PV that we assume to be independent of gain. As in the single-unit
case, the input to the first unit is a random variable Xs. with pdf PXs in the presence
of the signal and a random variable X A ? with pdf PXA in the absence of the signal. The
output of the first unit is described by the random variables YGS and YGA with pdfs PYas
and PYGA ? Now. because noise is added to the net input of the response unit as well. the
input of the response unit is the random variable Zas =YGS + V or ZGA = YGA + V. again
depending on whetber the signal is present or absent We write PZas and pz.ru for the
pdfs of these random variables. fJZos is the convolution of pyos and PV, and pz.ru is the
convolution of PYGA and Pv. The effect of convolving the output pdfs of the input unit
with the noise distribution is to increase the overlap between the resulting distributions
(PZas and pz.ru). and therefore decrease the discriminability of the input to the response
unit.
How are these distributions affected by an increase in gain on the input unit? By the
Constant Optimal Perfonnance Theorem. we already know that the overlap between PYGS
and PYGA remains constant as gain increases. Furthermore. as stated above, we have
assumed that the noise distribution is independent of gain. It would therefore seem that
a change in gain should not affect the overlap between PZos and pz.ru. However. it is
2We present the intuitions underlying our results in tenns of the overlap between the pdfs. However, the
proofs themselves are analytical.
The Effect of Catecholamines on Performance: From Unit to System Behavior
possible to show that. under very general conditions, the overlap between PZos and pz.a..
decreases when the gain of the input unit increases, thereby improving perfonnance of the
two-layered system. We call this the chain effect; the Chain Performance Theorem [10]
gives sufficient conditions for its appearance. 3
Paradoxically. the chain effect arises because the noise added to the net input of the
response unit is not affected by variations in the gain. As we mentioned before, increasing
the gain separates the means of the output pdfs of the input unit. I-'(YGS) and I-'(YGA)
(eventhough this does not affect the performance of the first unit). Suppose all the
probability mass were concentrated at these means. Then PZos would be a copy of Pv
centered at I-'(YGS). and pz.a.. would be a copy of pv centered at I-'(YGS). Thus in this case,
increasing the gain does correspond to rigidly translating PZos and PZat. apart, thereby
reducing their overlap and improving performance.
1
10
.
]
..
~
-4/J
?2/J
O/J
2/J
4/J
6/J
x(Ne' rnplll)
-4/J
?2/J
O/J
2/J
4/J
6/J
x (Nell""Ul)
---------p-~-------
-------p~------------
Figure 4: Dependence of Chain Output Pdfs Upon Gain. These graphs use the same conventions
and input pdfs as Figure 2. They depict the output pdfs, in the presence of additive Gaussian noise,
for G =0.5 (top) and G = 1.0 (bottom),
A similar effect arises in more general circumstances, when PYas and PY(JA are not concentrated at their means. Figure 4 provides an example. illustrating PZas and PZat. for
three different values of the gain. The first unit outputs are the same as in Figure 2, but
3In this discussion, we have assumed that the same noise was added to the net input into each unit of a
chain. However, the improvement in performance of a chain of units with increasing gain does not depend on
this particular assumption.
105
106
Servan-Schreiber, Printz and Cohen
these have been convolved with the pdf PV of a Gaussian random variable to obtain the
curves shown. Careful inspection of the figure will reveal that the overlap between PZa
and PZaA decreases as the gain rises.
4 Simulation of the Continuous Performance Test
The above analysis has shown that increasing the gain of the response function of individual units in a very simple network can improve signal detection performance. We
now present computer simulation results showing that this phenomenon may account for
improvements of performance with catecholamine agonists in a common behavioral test
of signal detection.
The continous performance test (CPT) has been used extensively to study attention and
vigilance in behavioral and clinical research. Performance on this task has been shown to
be sensitive to drugs or pathological conditions affecting catecholamine systems [11.8.9].
In this task, individual letters are displayed tachystoscopically in a sequence on a computer
monitor. In one common version of the task, a target event is to be reported when two
consecutive letters are identical. During baseline performance. subjects typically fail to
report 10 to 20% of targets ("misses") and inappropriately report a target during 0.5 to
1% of the remaining events ("false alarms"). Following the administration of agents
that directly release catecholamines from synaptic terminals and block re-uptake from
the synaptic cleft (i.e., CNS stimulants such as amphetamines or methylphenidate) the
number of misses decreases. while the number of false alarms remains approximately the
same. Using standard signal detection theory measures, investigators have claimed that
this pattern of results reflects an improvement in the discrimination between signal and
non-signal events (d'), while the response criterion (f3) does not vary significantly [11.8,9].
We used the backpropagation learning algorithm to train a recurrent three layer network
to perform the CPT (see Figure 5). In this model, several simplifyng assumptions made
in the preceding section are removed: in contrast to the simple two-unit assembly. the
network contains three layers of units (input layer, intermediate - or hidden - layer, and
output layer) with some recurrent connections; connection weights between these layers
are developed entirely by the training procedure; as a result, the activation patterns on
the intermediate layer that are evoked by the presence or absence of a signal are also
determined solely by the training procedure; finally. the representation of the signal is
distributed over an ensemble of units rather than determined by a single unit
Following training, Gaussian noise with zero mean was added to the net input of each unit
in the intermediate and output layers as each letter was presented. The overall standard
deviation of the noise distribution and the threshold of the response unit were adjusted
to produce a performance equivalent to that of subjects under baseline conditions (13.0%
misses and 0.75% false alarms). We then increased the gain of all the intermediate
and output units from 1.0 to 1.1 to simulate the effect of catecholamine release in the
network. This manipulation resulted in rates of 6.6% misses and 0.78% false alarms.
The correspondence between the network's behavior and empirical data is illustrated in
Figure 5.
The EfTect of Catecholamines on Performance: From Unit to System Behavior
Letter Identification Module
c~5
16 ......-
~~~
"r. .??~J
I \
~ .y
_
_ _ _ _ _ _ _ _ _ _......,
......
~F. . ........
--0-
Sim. ... _
_
6om.F._
;
J
t
ol-____.........
~O::::::::::~I~__--J
a.s
&lmoAonI
Feature Input Module
Figure 5: Simulation of the Continuous Performance Task. Len panel: The recurrent three-layer
network (12 input units, 30 intermediate units, 10 output units and 1 response unit). Each unit
projects to all units in the subsequent layer. In addition, each output unit also projects to each unit
in the intermediate layer. The gain parameter G is the same for all intermediate and output units.
In the simulation of the placebo condition, G = 1; in the simulation of the drug condition, G = 1.1.
The bias B = -1 in both conditions. Right panel: Performance of human subjects [9], and of the
simulation, on the CPT. With methylphenidate misses dropped from 11.7% to 5.5%, false alarms
decreased from 0.6% to 0.5% (non-significant).
The enhancement of signal detection performance in the simulation is a robust effect. It
appears when gain is increased in the intermediate layer only, in the output layer only,
or in both layers. Because of the recurrent connections between the output layer and
the intermediate layer, a chain effect occurs between these two layers when the gain is
increased over anyone of them, or both of them. The impact of the chain effect is to
reduce the distortion, due to internal noise, of the distributed representation on the layer
receiving inputs from the layer where gain is increased. Note also that the improvement
takes place even though there is no noise added to the input of the response unit. The
response unit in this network acts only as an indicator of the strength of the signal in
the intermediate layer. Finally, as the Constant Optimal Performance Theorem predicts,
increasing the gain only on the response unit does not affect the performance of the
network.
5 Conclusion
Fluctuations in catecholaminergic tone accompany psychological states such as arousal,
motivation and stress. Furthermore, dysfunctions of catecholamine systems are implicated
in several of the major psychiatric disorders. However, in the absence of models relating
changes in cell function to changes in system performance, the relation of catecholamines
to behavior has remained obscure. The findings reported in this paper suggest that the
behavioral impact of catecholamines depend on their effects on an ensemble of units
operating in the presence of noise, and not just on changes in individual unit responses.
107
108
Servan-Schreiber, Printz and Cohen
Furthermore. they indicate how neuromodulatory effects can be incorporated in parallel
distributed processing models of behavior.
References
[1] Y. Burnod and H. Korn. Consequences of stochastic release of neurotransmitters
for network computation in the central nervous system. Proceedings of the National
Academy of Science. 86:352-356. 1988.
[2] L. A. Chiodo and T. W. Berger. Interactions between dopamine and amino acid~
induced excitation and inhibition in the striatum. Brain Research. 375:198-203.
1986.
[3] C. R. Clark. G. M. Geffen, and L. B. Geffen. Catecholamines and attention ii:
pharmacological studies in normal humans. Neuroscience and Behavioral Reviews,
11:353-364, 1987.
[4] S. L. Foote. Extrathalamic modulation of cortical function. Ann. Rev. Neurosci.,
10:67-95, 1987.
[5] S. L. Foote, R. Freedman. and A. P. Olivier. Effects of putative neurotransmitters
on neuronal activity in monkey auditory cortex. Brain Research, 86:229-242, 1975.
[6] W. J. Freeman. Nonlinear gain mediating cortical stimulus-response relations. Biological Cybernetics, 33:243-247, 1979.
[7] G. E Hinton and Sejnowski T. J. Analyzing cooperative computation. Proceedings
of the Cognitive Science Society, 1983.
[8] R. K1orman, L. O. Bauer, H. W. Coons, J. L. Lewis, L. J. Peloquin, R. A. Perlmutter,
R. M. Ryan, L. F. Salzman, and J. Strauss. Enhancing effects of methylphenidate
on normal young adults cognitive processes. Psychopharmacology Bulletin, 20:3-9,
1984.
[9] L. J. Peloquin and R. K1orman. Effects of methylphenidate on normal children's
mood, event-related potentials, and performance in memory scanning and vigilance.
Journal of Abnormal Psychology, 95:88-98, 1986.
[10] H. Printz and D. Servan~Schreiber. Foundations of a Computational Theory of Catecholamine Effects. Technical Report CMU-CS-90~105. Carnegie Mellon, School
of Computer Science. 1990.
[11] J. Rapoport, M. S. Buchsbaum, H. Weingartner, T. P. Zahn. C. Ludlow, J. Bartko.
E. J. Mikkelsen, D. H. Langer, and Bunney W. E. Dextroamphetamine: cognitive
and behavioral effects in normal and hyperactive boys and normal adult males.
Archives of General Psychiatry. 37:933-943. 1980.
[12] M. Segal. Mechanisms of action of noradrenaline in the brain. Physiological Psychology, 13:172-178, 1985.
| 210 |@word illustrating:1 version:1 simulation:10 thereby:2 responsivity:8 contains:1 activation:9 nell:2 subsequent:1 additive:1 shape:2 motor:2 depict:1 discrimination:1 nervous:2 tone:3 inspection:1 provides:2 along:1 behavioral:9 expected:1 presumed:1 behavior:11 themselves:1 brain:4 terminal:1 ol:1 freeman:1 actual:2 increasing:13 psychopharmacology:1 project:2 underlying:1 maximizes:1 mass:1 panel:2 null:1 hyperactive:1 interpreted:1 monkey:1 developed:1 finding:1 act:1 wrong:1 hit:2 unit:66 uo:3 appear:1 positive:2 before:1 dropped:1 consequence:2 striatum:1 analyzing:1 rigidly:1 firing:1 solely:1 approximately:1 fluctuation:1 modulation:1 discriminability:1 evoked:1 block:1 geffen:2 backpropagation:1 procedure:2 area:1 empirical:1 drug:2 significantly:1 word:1 induce:1 spite:1 psychiatric:1 suggest:1 get:1 layered:1 coon:1 py:1 equivalent:1 attention:3 disorder:1 variation:1 target:4 suppose:1 olivier:1 pa:1 element:9 predicts:1 cooperative:1 observed:1 bottom:3 module:2 decrease:5 removed:1 mentioned:1 intuition:1 depend:2 solving:1 upon:1 neurotransmitter:2 train:1 describe:1 sejnowski:1 valued:1 distortion:1 ability:2 mood:1 sequence:2 differentiable:1 net:11 analytical:1 interaction:1 academy:1 enhancement:2 asymmetry:1 diverges:1 optimum:1 produce:2 converges:3 depending:1 recurrent:4 ij:1 school:2 sim:1 c:1 pxa:5 indicate:1 convention:1 closely:1 correct:2 stochastic:1 centered:2 human:4 translating:1 require:2 ja:1 biological:1 ryan:1 adjusted:1 strictly:4 noradrenaline:1 around:2 normal:5 mapping:1 major:1 vary:1 consecutive:1 sensitive:1 schreiber:7 reflects:1 catecholaminergic:3 gaussian:3 rather:2 release:5 improvement:7 pdfs:18 contrast:1 psychiatry:1 baseline:2 detect:2 entire:1 lj:1 typically:1 amphetamine:1 hidden:1 relation:3 overall:1 f3:1 identical:1 report:3 stimulus:3 primarily:1 pathological:1 resulted:1 national:1 individual:7 zahn:1 cns:1 detection:17 male:1 chain:14 closer:2 perfonnance:3 re:1 arousal:1 psychological:1 increased:5 instance:1 servan:7 deviation:1 placebo:1 reported:2 varies:1 scanning:1 combined:1 density:2 receiving:2 together:1 again:1 central:2 reflect:1 vigilance:2 fir:1 cognitive:3 convolving:1 account:5 potential:1 segal:1 de:1 harry:1 afferent:3 analyze:1 start:1 len:1 parallel:1 om:1 pya:5 acid:1 ensemble:2 correspond:3 yield:1 modelled:1 identification:1 agonist:1 cybernetics:1 explain:1 synaptic:3 steadily:2 associated:1 proof:1 gain:31 auditory:1 proved:1 appears:1 response:28 burnod:1 though:2 furthermore:3 xa:6 just:1 eventhough:1 receives:2 nonlinear:1 logistic:5 reveal:1 effect:29 illustrated:2 potentiating:1 during:2 dysfunction:1 pharmacological:1 excitation:1 criterion:1 pdf:5 stress:1 common:2 cohen:6 relating:1 mellon:2 significant:1 neuromodulatory:1 cortex:1 operating:1 inhibition:1 optimizes:1 apart:1 manipulation:1 claimed:1 tenns:1 preceding:1 determine:4 monotonically:1 signal:45 ii:1 relates:1 technical:1 clinical:1 long:1 post:1 impact:2 circumstance:1 enhancing:1 cmu:1 dopamine:3 represent:3 achieved:1 cell:7 affecting:1 addition:1 interval:1 decreased:1 printz:7 biased:1 archive:1 subject:5 induced:2 accompany:1 db:1 member:1 seem:1 call:4 presence:11 intermediate:10 paradoxically:1 affect:7 buchsbaum:1 psychology:3 reduce:1 administration:1 absent:4 motivated:1 ul:1 action:1 adequate:1 cpt:3 amount:2 extensively:2 concentrated:2 inhibitory:4 gil:1 neuroscience:1 arising:1 carnegie:2 discrete:1 write:1 affected:2 threshold:8 monitor:2 drawn:1 graph:4 langer:1 everywhere:2 letter:4 place:1 family:3 almost:2 putative:1 entirely:1 layer:20 abnormal:1 correspondence:1 activity:1 strength:2 precisely:1 simulate:1 anyone:1 performing:1 px:5 department:1 rev:1 lasting:1 pr:4 equation:1 remains:2 fail:1 mechanism:1 know:1 detennine:1 pdt:1 convolved:1 top:2 remaining:1 assembly:1 society:1 added:7 quantity:1 occurs:2 arrangement:1 already:1 imbedded:1 dependence:1 separate:1 argue:1 salzman:1 ru:4 berger:1 ratio:2 mediating:1 relate:1 boy:1 negative:3 stated:2 rise:1 perform:1 neuron:5 convolution:2 displayed:1 logistics:1 payoff:2 hinton:1 incorporated:1 david:1 continous:1 connection:3 cleft:1 adult:2 suggested:1 pattern:2 catecholamine:22 including:1 memory:2 oj:1 suitable:1 overlap:11 event:4 indicator:1 korn:1 improve:3 ne:1 axis:2 uptake:1 mediated:1 review:3 prior:1 analogy:1 ygs:8 clark:1 foundation:1 agent:1 sufficient:1 obscure:1 excitatory:4 copy:2 implicated:1 bias:4 foote:2 wide:1 bulletin:1 fg:2 distributed:5 bauer:1 curve:3 cortical:2 sensory:1 ignores:1 made:1 rapoport:1 approximate:1 ignore:1 pittsburgh:1 assumed:2 continuous:4 norepinephrine:2 tonoise:1 robust:1 improving:2 perlmutter:1 inappropriately:1 neurosci:1 whole:1 noise:18 alarm:8 motivation:1 freedman:1 child:1 convey:1 amino:1 neuronal:1 en:1 pv:6 young:1 theorem:4 remained:1 substance:1 showing:2 x:7 pz:6 physiological:1 consist:1 false:8 strauss:1 appearance:1 applies:1 satisfies:1 lewis:1 modulate:1 ann:1 careful:1 absence:10 change:21 typical:1 determined:2 reducing:1 miss:7 called:1 select:1 internal:1 arises:2 jonathan:1 categorize:1 investigator:2 phenomenon:2 |
1,208 | 2,100 | Efficiency versus Convergence of Boolean
Kernels for On-Line Learning Algorithms
Roni Khardon
Tufts University
Medford, MA 02155
[email protected]
Dan Roth
University of Illinois
Urbana, IL 61801
[email protected]
Rocco Servedio
Harvard University
Cambridge, MA 02138
[email protected]
Abstract
We study online learning in Boolean domains using kernels which capture feature expansions equivalent to using conjunctions over basic features. We demonstrate a tradeoff between the computational efficiency
with which these kernels can be computed and the generalization ability of the resulting classifier. We first describe several kernel functions
which capture either limited forms of conjunctions or all conjunctions.
We show that these kernels can be used to efficiently run the Perceptron algorithm over an exponential number of conjunctions; however we
also prove that using such kernels the Perceptron algorithm can make
an exponential number of mistakes even when learning simple functions. We also consider an analogous use of kernel functions to run the
multiplicative-update Winnow algorithm over an expanded feature space
of exponentially many conjunctions. While known upper bounds imply
that Winnow can learn DNF formulae with a polynomial mistake bound
in this setting, we prove that it is computationally hard to simulate Winnow?s behavior for learning DNF over such a feature set, and thus that
such kernel functions for Winnow are not efficiently computable.
1 Introduction
The Perceptron and Winnow algorithms are well known learning algorithms that make predictions using a linear function in their feature space. Despite their limited expressiveness,
they have been applied successfully in recent years to several large scale real world classification problems. The SNoW system [7, 2], for example, has successfully applied variations
of Perceptron [6] and Winnow [4] to problems in natural language processing. The system
first extracts Boolean features from examples (given as text) and then runs learning algorithms over restricted conjunctions of these basic features.
There are several ways to enhance the set of features after the initial extraction. One idea is
to expand the set of basic features using conjunctions such as
and use
these expanded higher-dimensional examples, in which each conjunction plays the role of
a basic feature, for learning. This approach clearly leads to an increase in expressiveness
and thus may improve
performance. However, it also dramatically increases the number of
features (from
to
if all conjunctions are used) and thus may adversely affect both the
computation time and convergence rate of learning.
This paper studies the computational efficiency and convergence of the Perceptron and
Winnow algorithms over such expanded feature spaces of conjunctions. Specifically, we
study the use of kernel functions to expand the feature space and thus enhance the learning abilities of Perceptron and Winnow; we refer to these enhanced algorithms as kernel
Perceptron and kernel Winnow.
1.1
Background: Perceptron and Winnow
Throughout its execution Perceptron maintains a weight vector
which is initially
Upon receiving an example
the algorithm predicts according to the
If the prediction is and the label is (false positive
linear threshold function
prediction) then the vector is set to
, while if the prediction is
and the label is
No change is made if the prediction is correct.
(false negative) then is set to
The famous Perceptron Convergence Theorem [6] bounds the number of mistakes which
the Perceptron algorithm can make:
( ') +* , -. /10
-1 "! / "for#all
,
2436; 583795:3
The Winnow algorithm [4] has a very similar structure.
Winnow maintains a hypothesis
vector <# which is initially >= BA Winnow is parameterized by a promotion
factor ? and a threshold @ 0
upon receiving an example C' +*D Winnow
predicts according to the threshold function ( @ If the prediction is and the label is
demotion step. If
then for all , such that =E the value of is set to GF ? ; =Hthis isthea value
of
is set
the prediction is and the label is then for all , such that
to ?I ; this is a promotion step. No change is made if the prediction is correct.
For our purposes the following mistake bound, implicit in [4], is of interest:
=
Theorem 2 Let the target function be a J -literal monotone disjunction K
in ' +*D labeled according
to
K the number
8LM NN M PO For any sequence of examples
of prediction mistakes made by Winnow ? @ is at most Q S J ?TU VXWZY\[ @
Q
QBR
$ $%&
be a sequence of labeled examples with
Theorem 1 Let
and
for all . Let
be such that
Then Perceptron makes at most
mistakes on this example sequence.
1.2
Our Results
Our first result in Section 2 shows that it is possible to efficiently run the kernel Perceptron
algorithm over an exponential number of conjunctive features:
]
]
Theorem 3 There is an algorithm that simulates Perceptron over the -dimensional feature space
of labeled examples
of all conjunctions of
basic features. Given a sequence
the prediction and update for each example take poly
time steps.
in
' ^*
This result is closely related to one of the main open problems in learning theory: efficient
learnability of disjunctions of conjunctions, or DNF (Disjunctive Normal Form) expres
is true
sions.1 Since linear threshold elements can represent disjunctions (e.g.
iff
), Theorems 1 and 3 imply that kernel Perceptron can be used to learn
DNF. However, in this framework the values of
and in Theorem 1 can be exponentially large, and hence the mistake bound given by Theorem 1 is exponential rather than
polynomial in
The question thus arises whether, for kernel Perceptron, the exponential
!_
`
&
M !_ M
1
Angluin [1] proved that DNF expressions cannot be learned efficiently using hypotheses which
are themselves DNF expressions from equivalence queries and thus also in the mistake bound model
which we are considering here. However this result does not preclude the efficient learnability of DNF
using a different class of hypotheses such as those generated by the kernel Perceptron algorithm.
upper bound implied by Theorem 1 is essentially tight. We give an affirmative answer, thus
showing that kernel Perceptron cannot efficiently learn DNF:
K
Theorem 4 There is a monotone DNF over and a sequence ofexamples
la
beled according to which causes the kernel Perceptron algorithm to make
mistakes.
K
@ =H` ? @
Turning to Winnow, an attractive feature of Theorem 2 is that for suitable the bound
is logarithmic in the total number of features
(e.g.
). Therefore, as
and
noted by several researchers [5], if a Winnow analogue of Theorem 3 could be obtained
this would imply efficient learnability of DNF. We show that no such analogue can exist:
`
?=
Theorem 5 There is no polynomial time algorithm which simulates Winnow over exponentially many monotone conjunctive features for learning monotone DNF, unless every
problem in #P can be solved in polynomial time.
We observe that, in contrast to Theorem 5, Maass and Warmuth have shown that the Winnow algorithm can be simulated efficiently over exponentially many conjunctive features
for learning some simple geometric concept classes [5].
While several of our results are negative, in practice one can achieve good performance
by using kernel Perceptron (if
is small) or the limited-conjunction kernel described in
Section 2 (if
is large). This is similar to common practice with polynomial kernels 2 where
typically a small degree is used to aid convergence. These observations are supported by
our preliminary experiments in an NLP domain which are not reported here.
2 Theorem 3: Kernel Perceptron with Exponentially Many Features
It is easily observed, and well known, that the hypothesis of the Perceptron algorithm
is
made. If we let
a sum of the previous examples on which prediction mistakes were
denote the label of example , then
where is the
set of examples on which the algorithm
made
a
mistake.
Thus
the
prediction
of Perceptron
on is 1 iff
.
"' ^ *
=
U =
B
=
For an example ' + * let
denote its transformation into an enhanced feature
space such as the space of all conjunctions.
To run
the Perceptron algorithm over the
where
is the
enhanced space we must predict iff
B
weight
B
vector
in the .
enhanced space;
from
the
above
discussion
this
holds
iff
Denoting
= B
this holds iff
.
Thus we never need to construct the enhanced
feature space explicitly; we need only be
able to compute the kernel function
efficiently. This is the idea behind all so-called
kernel methods, which can be applied to any algorithm (such as support vector machines)
whose prediction is a function of inner products of examples; see e.g. [3] for a discussion.
The result in Theorem 3 is simply obtained by presenting a kernel function capturing all
conjunctions. We also describe kernels for all monotone conjunctions which allow no
negative literals, and kernels capturing all (monotone) conjunctions of up to literals.
J
The general
includes all conjunctions (with positive and negative
case:
mustWhen
literals)
compute the number of conjunctions which are true in both and
. Clearly, any literal in such a conjunction must satisfy both and and thus the corresponding
must have the
=bit"!$in# &%' ( where
same
value. Counting all such conjunctions gives
)+*-,/. is the number of original features that have the
same value in and . This kernel has been obtained independently by [8].
2
Our Boolean kernels are different than standard polynomial kernels in that all the conjunctions
are weighted equally. While expressive power does not change, convergence and behavior, do.
Monotone Monomials: In some applications the total number
of basic features may
be very large but in any one example only a small number of features take value 1. In
other applications the number of features
may not be known in advance (e.g. due to
unseen words in text domains). In these cases it may be useful to consider only&%monotone
'(
+! #
monomials. To express all monotone monomials we take
where
)+*,/. ) is the number of active features common to both and .
=
!Y
A parameterized kernel: In general, one may want to trade off expressivity against
number of examples and convergence time. Thus we consider a parameterized kernel
which captures all conjunctions of size at most for some
&%The
' ( number of
. This kersuch conjunctions that satisfy both and is
"!$#
nel is reported also in [10].
For
monotone
conjunctions
of
size
at
most
we have
&%' (
.
+! #
=
J
=
J
J
3 Theorem 4: Kernel Perceptron with Exponentially Many Mistakes
We describe a monotone DNF target function and a sequence of labeled examples which
cause the monotone kernel Perceptron algorithm to make exponentially many mistakes.
C'
+* we write to denote the number of 1?s in and . to denote
)+*,/.!Y)
We use the following lemma (constants have not been optimized):
-bit
*
% ' +* % with ]
=
strings =E' %
Lemma 6 There is a set of
%
%
such that \=
F
for
, ] and ! #"
F%$ for ,&(' ]
Proof: The proof uses the probabilistic method. For each ,V= ] let ! X' +* be
probability 1/10.
chosen by independently
A setting each bit to with
- For
% Rany2 , itandis clear
that )+*, - =
F a Chernoff bound implies
that
thus
.0/%*1 !
2 F
Similarly, for any ,=6
the probability that any satisfies
is
at
most
3
4
]
5 %'
R
F
7 #" -#=
DBA a Chernoff bound implies that .8/%*1 9 #" 0
+
)
1
*
wehave
F
F
:
$
2
and thus the probability that any " with ;
R
,
=
6
5 2 ' satisfies
" 0
%
F
$ is at
<
2
3
<
2
4] R % is less than 1. Thus
For ] =
most _ R
the value of _ R
=
F and # #"
:F $ For any
we have each
for some choice of
0
which has
F we can set ^
F of the 1s to 0s, and the lemma is proved.
The target DNF is very simple: it is the single conjunction !_ While
the original
Perceptron algorithm over the
features makes at most poly
mistakes for
this target function,
we now show that the monotone kernel
Perceptron algorithm which
runs over all
monotone monomials can make >
mistakes.
Recall that at the beginning of the Perceptron algorithm?s
execution all
A since
= coordinates
Perceptronof
are 0. The first example is the negative example
incorrectly predicts 1 on this example. The resulting update causes the coefficient ?
corresponding
to the empty monomial (satisfied by any example ) to become but all
other coordinates of remain
0. The next example is the positive example
For this
=
so Perceptron incorrectly predicts Since
example we have
all
monotone conjunctions
are satisfied by this example the resulting update causes
? to become 0 and all
other
coordinates of to become 1. The next
described in Lemma 6. Since each such example has
examples are
the vectors
)=
F each
example is negative; however as we now show the Perceptron algorithm
will predict on each of these examples.
% , % and consider the hypothesis vector just before
Fix any value
example
is received. Since !=
F the value of
is a sum of the _ different
For
!
coordinates which correspond to the monomials satisfied by More precisely we
where contains the monomials which are
have
and #" for some '65
and contains the monomials which are satisfied
satisfied by
by but no #" with '5 We lower bound the two sums separately.
Let
be any monomial in By
contains at most
%$ variables
2 Lemma
6 any
monomials in Using the well known bound
and thus there can be
at
most
"
; "
and
is the binary entropy function there can be
where
34
at most
terms in
Moreover the value of each must be at least34 since
decreases by at most 1 for each example, and hence
;
By Lemma 6 for
On the other hand, for any
we clearly have
:$ every
any
satisfied by
must belong to and hence
3 -variable
monomial
Combining these inequalities we have
< 2
and hence the Perceptron prediction on is 1.
=
%
Q
Q
=,
)
=E,
_
? % F
F
H
?
=
_
0 F _
_
0
_
0
0
^*
.
'
are all nonempty monomials
conjunctions) over A sequence of la (monotone
consistent if it is consistent with some
beled examples
% " for isallmonotone
implies % " If is monotone
monotone function, i.e.
J
=
consistent and has ] labeled examples then clearly there is a monotone DNF formula consistent with which contains at most ] conjunctions. We consider the following problem:
KERNEL WINNOW PREDICTION ? @ (KWP)
of labeled examples
Instance: Monotone
consistent sequence C= A example
with each .' ^ *
and each .') +* unlabeled
' +*
Question: Is
vector
@ where
is the ` = . -dimensional
hypothesis
?
generated by running Winnow ? @ on the example sequence
In order to run Winnow over all
. nonempty monomials to learn monotone DNF, one
must be able to solve KWP efficiently. The main result of this section is proved by showing
4 Theorem 5: Learning DNF with Kernel Winnow is Hard
In this section, for
denotes the
-element vector whose coordinates
that KWP is computationally hard for any parameter settings which yield a polynomial
mistake bound for Winnow via Theorem 2.
Q S ? >
`
=
U and ? 0 @ be such that , *
QBR
WZY\[ Q @
=
Then KWP ? @
is #P-hard.
Proof of
it can easily be verified that
Theorem 7: For
! `
? and_#" @ as described above
? poly and @ The proof of the theorem is a
Theorem 7 Let
!
poly
poly
poly
reduction from the following #P-hard problem [9]: (See [9] also for details on #P.)
poly
= 8L4M 3
' +* ?
%_
= %
MONOTONE 2-SAT (M2SAT)
Instance: Monotone 2-CNF Boolean formula $ &% (' % )' *' % with %
integer such that
and each ,+
Question: Is $
i.e. does $ have at least satisfying assignments in
'
* A
R
4.1
High-Level Idea of the Proof
The high level idea of the proof is simple: let $
be an instance of M2SAT where
$ is defined over variables The Winnow algorithm maintains a weight
for each monomial
over variables We define
a 1-1 correspondence between
for $ and we give a sequence of
these monomials
and truth assignments
examples for Winnow which causes .if $
and
if $
' +*
=
=
=
A
some additional work ensures that
R
@
R
W YD[ ? =
=
_
In more detail, let =
_ WZYD[
Q
Q
Q
and
=
C
We describe a polynomial time transformation
which maps
an
-variable instance $
of M2SAT to an -variable instance of
KWP
consistent, each and belong
is monotone
? @ where X=E
to ' ^* and @ if and only if $ R
The Winnow variables
are divided into three sets and where
=
* = ' * and = ' * The unlabeled ex'ample
in and
are set to 1 and all variables in
is
R R
i.e. all variables
are set to 0. We thus have
= ( where = ?
= ?
and (
= '
? '
? We refer to monomials
= 5
as typemonomials,
= 5
as type- monomials, and
!
= 5 as type- monomials.
= 5
monomials
monomials
A
The example sequence is divided into four stages. Stage 1 results in - $ R as
described below the
variables in
correspond to the
variables in the CNF formula $
Stage 2 results in -
? " $ R for some final
positive integer # Stages
3 and 4 together
result in
(
- @ ? " Thus the
value
of
is
@# ? " $ R ^ so we have
@ if and only if $ R
approximately
Since all variables in
are 0 in if
includes a variable in then the value of
does not affect The variables in are ?slack variables? which (i) make Winnow
perform the correct promotions/demotions and (ii) ensure that is monotone consistent.
The value of is thus related to $
if and only if $
4.2
Details of the Proof
$& %('* ),+.-0/214365)
' +*
7 &8 =
= LM 3
8L = 3 =
=
= < % ? R_9
=
= R ;:
!:
? R9
We now show how the Stage 1 examples cause Winnow
to make a false positive predic and
for all other , in
tion on negative examples which have 8L =
=
=
as described above. For each such negative example
new slack variables
@ F 1
six repeated
=< =< ! are used as follows: Stage 13 hasinPW YDStage
instances of
[
the positive example which has >< % = =< _ = and all Q other bits 0. These examples
%@ ?BA L %@?BA %@?BA L %@?BA ? @ and hence
cause promotions which result in @
@
%
B
?
A
3 first with ><3
= =< =
groups of similar examples (the
L
@ F Two=< other
=<
C = = ) cause %@?BAD @ F and %@?BAE @ F The
the second with
next example in is the negative example which has 8L =
=
= example
for
=<
=<
=<
C
3
all other
in
and
all
other
bits
0.
For
this
=
=
=
0 %@ ?BA L %@ ?BAD (%@?BAE @ so Winnow makes a false positive prediction.
_
Since $ has at most
clauses and there
are negative examples per clause, this con_
struction can be carried out using
slack variables
9 3
F IHJ ) +. -0 /6 1K365) . The
Stage 2: Setting $&%'G
first Stage 2 example is a positive example
with
= for all ,
9 3 = and all other bits 0. Since each of the
Stage 1: Setting
. We define the following correspondence between
and monomials
truth assignments
if and only if
is not
present in
For each clause
in $ Stage 1 contains negative examples such
Assuming that (1) Winnow makes a
that
and
for all other
false positive prediction on each of these examples and (2) in Stage 1 Winnow never does a
promotion on any example which has any variable in set to 1, then after Stage 1 we will
Thus we will have
have that
if $
and
if $
for some
$
=
= R
9
6 ;3 :
= ? @ 0
: F
U ?
0
R
%
#=PW YD[ Q @ F
? " U@ ? "
#
= ?" R
N; :
"?
" @
N;:
R
?"
=>? " R N : U? " U@ F
$ '
=
=
=
? "
6 :
R
.
.
=>@) ? " R
@B
? @ " :
9
R R
: _
%
@
.? "
U@
( .? " _
@
R
For ease of notation let denote @9
the examples
? " We now
% describe
E _ ? "
in Stages 3 and 4 and show that they will cause
to satisfy
so ? " R % ? " and hence there is a unique smallest integer such that
Let %
= PW YD[
% ? " R Q ? " The Stage% 3 examples
will result in
% Using the
definition
it can be verified
we? " R have
the % fact that
? that ? " R
Hence
H % ? " of% @ and
?
?
?
H
=
?
"
"
"
"
R
R
? %
Q
=
We use the following lemma:
% % there is a monotone CNF $ ' over
Lemma 8 For all for all
' Boolean
variables which has at most clauses, has exactly satisfying assignments in
+* and can be constructed from and in poly
time.
Proof: The proof is by induction on . For the base case (= we have = and
Assuming the lemma is true for =< J we now prove it for = J 8
$ ' =
% % then the desired CNF is $ ' = ' $ -' Since $ -' has at most J
If
% % then the desired CNF
"
clauses $ ' has at most J " clauses. If
M $ -' _ O By distributing over each clause of $ -' _ O we can write
is $ ' =
R
'
as a CNF with at most J clauses. If T
$
= then $ ' = R
and are satisfied by this example have
monomials which contain
!
$
poly
we have -
after
Since
$
the resulting promotion we have
$
so
Let
Stage 2 consists of repeated
instances of the
described
above. After these promotions
we have
positive
example
Since $
we also
$
have
$
Stage 3: Setting
. At the start of Stage 3 each type- and type- monomial
There are
variables in and variables in so at the start of Stage 2
has
Since no example
we have
and (
in Stages 3 or 4
and (
satisfies any in
at the end of Stage 4
will still be
$
will still
Thus ideally at the end of Stage 4 the value of would
be
be
since this would imply that
$
for
which is at least if and only if $
However it is not necessary
as long as
must be an integer and
to assume
this exact value; since $
we will have
(
that
if and only if $
Let $ ' be an -clause monotone CNF formula over the variables in which has
satisfying assignments. Similar to Stage 1, for each clause of $ ' , Stage 3 has
negative
examples corresponding to that clause, and as in Stage 1 slack variables in are used to
ensure that Winnow makes a false positive prediction on each such negative example.
Thus
the examples in Stage 3 cause
where
Since six
slack variables in are used for each negative example and there are
negative
examples, the slack variables
are sufficient for Stage 3.
Stage 4: Setting
. All that remains is to perform &%
promotions on examples which have each in set to 1. This will cause
which is as desired.
:_
R_
=
:_
_
9
3
$ $7% ' FIH
;: _
? " R = C ? " : _ ? " R E _ ? "
? R %
% ? " #1R
#
R Q
=
.: _
R Q
% The first
and % that
It can be verified from the definitions of
examples in are all the same positive example which has each
in set to 1 and
The first time this example is received
It can
R =<
performs a promotion; after #
R
occurrences
" @
so Winnow
Q%
_
"
R
of this example -
=
<
?
:
? " R
?
>@
"
_
and
=>? " R
;:
The remaining examples in Stage 4 are
R
% repetitions of the positive example
Q
which has each in set to 1 and
=
If promotions
on each repetition of
R occurred
" R
: _
so
this example then we would have
=
?
?
R
we need
show that this is less than @ We reexpress this quantity as ?
? " R Ionly
: _
We have ? " R I : _
? " R % ? _ " R % @
? " ? " U@9 _ ? "
Some
R
Q_
R _ ? " so indeed U@
easy manipulations show that ?
be verified that
? @=
Finally, we observe that by construction the example sequence is monotone consistent.
Since
poly
and contains poly
examples the transformation from M2SAT to
KWP
is polynomial-time computable and the theorem is proved.
(Theorem 7)
5 Conclusion
It is necessary to expand the feature space if linear learning algorithms are to learn expressive functions. This work explores the tradeoff between computational efficiency and
convergence (i.e. generalization ability) when using expanded feature spaces. We have
shown that additive and multiplicative update algorithms differ significantly in this respect;
we believe that this fact could have significant practical implications. Future directions
include the utilization of the kernels developed here and studying convergence issues of
Boolean-kernel Perceptron and Support Vector Machines in the PAC model.
Acknowledgements: R. Khardon was supported by NSF grant IIS-0099446. D. Roth
was supported by NSF grants ITR-IIS-00-85836 and IIS-9984168 and by EPSRC grant
GR/N03167 while visiting University of Edinburgh. R. Servedio was supported by NSF
grant CCR-98-77049 and by a NSF Mathematical Sciences Postdoctoral Fellowship.
References
[1] D. Angluin. Negative results for equivalence queries. Machine Learning, 2:121?150, 1990.
[2] A. Carlson, C. Cumby, J. Rosen, and D. Roth. The SNoW learning architecture. Technical
Report UIUCDCS-R-99-2101, UIUC Computer Science Department, May 1999.
[3] N. Cristianini and J. Shaw-Taylor. An Introduction to Support Vector Machines. Cambridge
Press, 2000.
[4] N. Littlestone. Learning quickly when irrelevant attributes abound: A new linear-threshold
algorithm. Machine Learning, 2:285?318, 1988.
[5] W. Maass and M. K. Warmuth. Efficient learning with virtual threshold gates. Information and
Computation, 141(1):378?386, 1998.
[6] A. Novikoff. On convergence proofs for perceptrons. In Proceeding of the Symposium on the
Mathematical Theory of Automata, volume 12, pages 615?622, 1963.
[7] D. Roth. Learning to resolve natural language ambiguities: A unified approach. In Proc. of the
American Association of Artificial Intelligence, pages 806?813, 1998.
[8] K. Sadohara. Learning of boolean functions using support vector machines. In Proc. of the
Conference on Algorithmic Learning Theory, pages 106?118. Springer, 2001. LNAI 2225.
[9] L. G. Valiant. The complexity of enumeration and reliability problems. SIAM Journal of Computing, 8:410?421, 1979.
[10] C. Watkins. Kernels from matching operations. Technical Report CSD-TR-98-07, Computer
Science Department, Royal Holloway, University of London, 1999.
| 2100 |@word pw:3 polynomial:9 open:1 tr:1 reduction:1 initial:1 contains:6 denoting:1 conjunctive:3 must:7 additive:1 update:5 intelligence:1 warmuth:2 beginning:1 mathematical:2 constructed:1 become:3 symposium:1 prove:3 consists:1 dan:1 indeed:1 behavior:2 themselves:1 uiuc:2 beled:2 resolve:1 enumeration:1 preclude:1 considering:1 struction:1 abound:1 moreover:1 notation:1 string:1 affirmative:1 developed:1 unified:1 transformation:3 every:2 exactly:1 classifier:1 utilization:1 grant:4 positive:13 before:1 mistake:16 despite:1 yd:3 approximately:1 equivalence:2 ease:1 limited:3 unique:1 practical:1 practice:2 hthis:1 significantly:1 matching:1 word:1 cannot:2 unlabeled:2 equivalent:1 map:1 roth:4 independently:2 automaton:1 variation:1 coordinate:5 analogous:1 enhanced:5 play:1 target:4 construction:1 exact:1 us:1 hypothesis:6 harvard:2 element:2 satisfying:3 predicts:4 labeled:6 observed:1 role:1 disjunctive:1 epsrc:1 solved:1 capture:3 ensures:1 trade:1 decrease:1 complexity:1 ideally:1 cristianini:1 tight:1 upon:2 efficiency:4 po:1 easily:2 describe:5 dnf:16 london:1 query:2 artificial:1 disjunction:3 whose:2 solve:1 ability:3 unseen:1 final:1 online:1 sequence:12 product:1 combining:1 iff:5 achieve:1 convergence:10 empty:1 received:2 wzy:1 c:1 implies:3 differ:1 direction:1 snow:2 closely:1 correct:3 attribute:1 virtual:1 fix:1 generalization:2 preliminary:1 hold:2 normal:1 algorithmic:1 predict:2 lm:1 smallest:1 purpose:1 proc:2 label:5 repetition:2 successfully:2 weighted:1 promotion:10 clearly:4 rather:1 r_:1 sion:1 conjunction:29 contrast:1 typically:1 lnai:1 initially:2 expand:3 issue:1 classification:1 construct:1 never:2 extraction:1 chernoff:2 future:1 rosen:1 report:2 novikoff:1 danr:1 interest:1 behind:1 ihj:1 implication:1 necessary:2 unless:1 taylor:1 littlestone:1 desired:3 instance:7 boolean:8 predic:1 assignment:5 monomials:18 gr:1 learnability:3 reported:2 answer:1 eec:1 explores:1 siam:1 probabilistic:1 off:1 receiving:2 enhance:2 together:1 quickly:1 ambiguity:1 satisfied:7 literal:5 adversely:1 american:1 includes:2 coefficient:1 satisfy:3 explicitly:1 multiplicative:2 tion:1 start:2 maintains:3 il:1 efficiently:8 correspond:2 yield:1 famous:1 researcher:1 expres:1 definition:2 against:1 servedio:2 proof:10 proved:4 recall:1 fih:1 higher:1 just:1 implicit:1 stage:28 hand:1 expressive:2 believe:1 concept:1 true:3 contain:1 hence:7 maass:2 attractive:1 noted:1 q_:1 presenting:1 demonstrate:1 performs:1 common:2 clause:11 exponentially:7 volume:1 belong:2 occurred:1 association:1 refer:2 significant:1 cambridge:2 similarly:1 illinois:1 language:2 reliability:1 base:1 xwzy:1 recent:1 winnow:34 irrelevant:1 manipulation:1 kwp:6 inequality:1 binary:1 additional:1 ii:4 technical:2 long:1 divided:2 equally:1 prediction:18 basic:6 essentially:1 kernel:39 represent:1 background:1 want:1 separately:1 fellowship:1 simulates:2 ample:1 integer:4 reexpress:1 counting:1 easy:1 affect:2 architecture:1 inner:1 idea:4 itr:1 tradeoff:2 computable:2 whether:1 expression:2 six:2 distributing:1 roni:2 cause:11 cnf:7 dramatically:1 useful:1 clear:1 nel:1 angluin:2 exist:1 nsf:4 per:1 ccr:1 demotion:2 write:2 express:1 group:1 four:1 threshold:6 verified:4 monotone:28 year:1 sum:3 run:7 parameterized:3 throughout:1 bit:6 capturing:2 bound:13 correspondence:2 precisely:1 simulate:1 expanded:4 department:2 according:4 remain:1 restricted:1 computationally:2 remains:1 slack:6 nonempty:2 end:2 studying:1 operation:1 observe:2 medford:1 tuft:2 occurrence:1 shaw:1 gate:1 original:2 denotes:1 running:1 nlp:1 ensure:2 remaining:1 include:1 carlson:1 implied:1 question:3 quantity:1 rocco:2 visiting:1 simulated:1 induction:1 assuming:2 negative:16 ba:11 perform:2 upper:2 observation:1 urbana:1 incorrectly:2 expressiveness:2 optimized:1 learned:1 expressivity:1 able:2 below:1 royal:1 analogue:2 power:1 suitable:1 natural:2 turning:1 improve:1 imply:4 carried:1 extract:1 gf:1 text:2 geometric:1 acknowledgement:1 versus:1 degree:1 sufficient:1 consistent:8 supported:4 monomial:5 allow:1 perceptron:33 edinburgh:1 world:1 made:5 active:1 sat:1 postdoctoral:1 learn:5 dea:1 expansion:1 poly:11 domain:3 main:2 csd:1 repeated:2 aid:1 khardon:2 exponential:5 watkins:1 formula:5 theorem:23 showing:2 pac:1 false:6 valiant:1 execution:2 entropy:1 logarithmic:1 simply:1 springer:1 truth:2 satisfies:3 ma:2 uiucdcs:1 hard:5 change:3 specifically:1 lemma:9 total:2 called:1 la:2 cumby:1 perceptrons:1 holloway:1 support:4 arises:1 ex:1 |
1,209 | 2,101 | An Efficient, Exact Algorithm for Solving
Tree-Structured Graphical Games
Michael L. Littman
AT&T Labs- Research
Florham Park, NJ 07932-0971
mlittman?research.att.com
Michael Kearns
Department of Computer & Information Science
University of Pennsylvania
Philadelphia, PA 19104-6389
mkearns?cis.upenn.edu
Satinder Singh
Syntek Capital
New York, NY 10019-4460
baveja?cs. colorado. edu
Abstract
We describe a new algorithm for computing a Nash equilibrium in
graphical games, a compact representation for multi-agent systems
that we introduced in previous work. The algorithm is the first
to compute equilibria both efficiently and exactly for a non-trivial
class of graphical games.
1
Introduction
Seeking to replicate the representational and computational benefits that graphical models have provided to probabilistic inference, several recent works
have introduced graph-theoretic frameworks for the study of multi-agent systems (La Mura 2000; Koller and Milch 2001; Kearns et al. 2001). In the simplest
of these formalisms, each vertex represents a single agent, and the edges represent
pairwise interaction between agents. As with many familiar network models, the
macroscopic behavior of a large system is thus implicitly described by its local interactions, and the computational challenge is to extract the global states of interest.
Classical game theory is typically used to model multi-agent interactions, and the
global states of interest are thus the so-called Nash equilibria, in which no agent
has a unilateral incentive to deviate.
In a recent paper (Kearns et al. 2001), we introduced such a graphical formalism for
multi-agent game theory, and provided two algorithms for computing Nash equilibria when the underlying graph is a tree (or is sufficiently sparse). The first algorithm
computes approximations to all Nash equilibria, in time polynomial in the size of
the representation and the quality of the desired approximation. A second and
related algorithm computes all Nash equilibria exactly, but in time exponential in
the number of agents. We thus left open the problem of efficiently computing exact
equilibria in sparse graphs.
In this paper, we describe a new algorithm that solves this problem. Given as input
a graphical game that is a tree, the algorithm computes in polynomial time an exact Nash equilibrium for the global multi-agent system. The main advances involve
the definition of a new data structure for representing "upstream" or partial Nash
equilibria, and a proof that this data structure can always be extended to a global
equilibrium. The new algorithm can also be extended to efficiently accommodate
parametric representations of the local game matrices, which are analogous to parametric conditional probability tables (such as noisy-OR and sigmoids) in Bayesian
networks.
The analogy between graphical models for multi-agent systems and probabilistic
inference is tempting and useful to an extent. The problem of computing Nash
equilibria in a graphical game, however, appears to be considerably more difficult
than computing conditional probabilities in Bayesian networks. Nevertheless, the
analogy and the work presented here suggest a number of interesting avenues for
further work in the intersection of game theory, network models, probabilistic inference, statistical physics, and other fields.
The paper is organized as follows. Section 2 introduces graphical games and other
necessary notation and definitions. Section 3 presents our algorithm and its analysis ,
and Section 4 gives a brief conclusion.
2
Preliminaries
An n-player, two-action 1 game is defined by a set of n matrices Mi (1 ~ i ~ n),
each with n indices. The entry Mi(Xl, ... ,xn ) = Mi(X) specifies the payoff to player
i when the joint action of the n players is x E {O, I} n. Thus, each Mi has 2n entries.
If a game is given by simply listing the 2n entries of each of the n matrices, we will
say that it is represented in tabular form.
?
The actions and 1 are the pure strategies of each player, while a mixed strategy
for player i is given by the probability Pi E [0, 1] that the player will play 1. For
any joint mixed strategy, given by a product distribution p, we define the expected
payoff to player i as Mi(i/) = Ex~p[Mi(X)], where x'" pindicates that each Xj is 1
with probability Pj and with probability 1 - Pj.
?
We use p[i : P:] to denote the vector that is the same as p except in the ith
component, where the value has been changed to P:. A Nash equilibrium for the
game is a mixed strategy p such that for any player i, and for any value
E
[0,1], Mi(i/) ::::: Mi(p[i :
(We say that Pi is a best response to jJ.) In other
words, no player can improve its expected payoff by deviating unilaterally from a
Nash equilibrium. The classic theorem of Nash (1951) states that for any game,
there exists a Nash equilibrium in the space of joint mixed strategies (product
distri butions).
pm.
P:
An n-player graphical game is a pair (G, M), where G is an undirected graph 2 on n
1 At present , no polynomial-time algorithm is known for finding Nash equilibria even in
2-player games with more than two actions, so we leave the extension of our work to the
multi-action setting for future work.
2The directed tree-structured case is trivial and is not addressed in this paper.
vertices and M is a set of n matrices Mi (1 ::; i ::; n), called the local game matrices .
Player i is represented by a vertex labeled i in G. We use N G (i) ~ {I, ... , n}
to denote the set of neighbors of player i in G- those vertices j such that the
undirected edge (i , j) appears in G. By convention, NG(i) always includes i itself.
The interpretation is that each player is in a game with only his neighbors in G.
Thus, if ING(i) I = k, the matrix Mi has k indices, one for each player in NG(i) , and
if x E [0, Ilk, Mi(X) denotes the payoff to i when his k neighbors (which include
himself) play
The expected payoff under a mixed strategy jJ E [0, Ilk is defined
analogously. Note that in the two-action case, Mi has 2k entries, which may be
considerably smaller than 2n.
x.
Since we identify players with vertices in G, it will be easier to treat vertices symbolically (such as U, V and W) rather than by integer indices. We thus use Mv to
denote the local game matrix for the player identified with vertex V.
Note that our definitions are entirely representational, and alter nothing about the
underlying game theory. Thus, every graphical game has a Nash equilibrium. Furthermore, every game can be trivially represented as a graphical game by choosing
G to be the complete graph and letting the local game matrices be the original
tabular form matrices. Indeed, in some cases, this may be the most compact graphical representation of the tabular game. However, exactly as for Bayesian networks
and other graphical models for probabilistic inference, any game in which the local
neighborhoods in G can be bounded by k ? n, exponential space savings accrue.
The algorithm presented here demonstrates that for trees, exponential computational benefits may also be realized.
3
The Algorithm
If (G , M) is a graphical game in which G is a tree, then we can always designate
some vertex Z as the root. For any vertex V, the single neighbor of Von the path
from V to Z shall be called the child of V, and the (possibly many) neighbors of V
on paths towards the leaves shall be called the parents of V. Our algorithm consists
of two passes: a downstream pass in which local data structures are passed from the
leaves towards the root, and an upstream pass progressing from the root towards
the leaves.
Throughout the ensuing discussion, we consider a fixed vertex V with parents
U I , ... , Uk and child W. On the downstream pass of our algorithm, vertex V will
compute and pass to its child W a breakpoint policy, which we now define.
?
Definition 1 A breakpoint policy for V consists of an ordered set of W -breakpoints
=
< WI < W2 < ... < Wt-I < Wt = 1 and an associated set of V-values
VI , . .. ,Vt? The interpretation is that for any W E [0,1], if Wi-I < W < Wi for some
index i and W plays w, then V shall play Vii and if W = Wi for some index i , then
V shall play any value between Vi and Vi+I. We say such a breakpoint policy has
t - 1 breakpoints.
Wo
A breakpoint policy for V can thus be seen as assigning a value (or range of values)
to the mixed strategy played by V in response to the play of its child W. In a slight
abuse of notation, we will denote this breakpoint policy as a function Fv(w), with
the understanding that the assignment V = Fv(w) means that V plays either the
fixed value determined by the breakpoint policy (in the case that W falls between
breakpoints), or plays any value in the interval determined by the breakpoint policy
(in the case that W equals some breakpoint).
Let G V denote the subtree of G with root V, and let M~=w denote the subset
of the set of local game matrices M corresponding to the vertices in G V , except
that the matrix M v is collapsed one index by setting W = w, thus marginalizing
W out. On its downstream pass, our algorithm shall maintain the invariant that if
we set the child W = w, then there is a Nash equilibrium for the graphical game
(G v , M~=w) (an upstream Nash) in which V = Fv(w). If this property is satisfied
by Fv(w), we shall say that Fv(w) is a Nash breakpoint policy for V. Note that
since (Gv, M~=w) is just another graphical game, it of course has (perhaps many)
Nash equilibria, and V is assigned some value in each. The trick is to commit
to one of these values (as specified by Fv (w)) that can be extended to a Nash
equilibrium for the entire tree G, before we have even processed the tree below V .
Accomplishing this efficiently and exactly is one of the main advances in this work
over our previous algorithm (Kearns et al. 2001).
The algorithm and analysis are inductive: V computes a Nash breakpoint policy
Fv(w) from Nash breakpoint policies FUl (v), ... , FUk (v) passed down from its parents (and from the local game matrix Mv). The complexity analysis bounds the
number of breakpoints for any vertex in the tree. We now describe the inductive
step and its analysis.
3.1
Downstream Pass
For any setting it E [0, l]k for
-0 and w
~v(i1,w)
E [0,1] for W, let us define
== Mv(l,it,w) - Mv(O,it,w).
The sign of ~v(it, w) tells us V's best response to the setting of the local neighborhood -0 = it, W = w; positive sign means V = 1 is the best response, negative that
V = 0 is the best response, and 0 that V is indifferent and may play any mixed
strategy. Note also that we can express ~v(it,w) as a linear function of w:
~v(it,w) = ~v(it, O)
+ w(~v(it, 1) -
~v(it,
0)).
For the base case, suppose V is a leaf with child W; we want to describe the Nash
breakpoint policy for V. If for all w E [0,1], the function ~v(w) is non-negative
(non-positive, respectively), V can choose 1 (0, respectively) as a best response
(which in this base case is an upstream Nash) to all values W = w. Otherwise,
~ v (w) crosses the w-axis, separating the values of w for which V should choose
1, 0, or be indifferent (at the crossing point). Thus, this crossing point becomes
the single breakpoint in Fv(w). Note that if V is indifferent for all values of w, we
assume without loss of generality that V plays l.
The following theorem is the centerpiece of the analysis.
Theorem 2 Let vertex V have parents UI , ... ,Uk and child W, and assume V has
received Nash breakpoint policies FUi (v) from each parent Ui . Then V can efficiently
compute a Nash breakpoint policy Fv (w). The number of breakpoints is no more
than two plus the total number of breakpoints in the FUi (v) policies.
Proof: Recall that for any fixed value of v, the breakpoint policy FUi (v) specifies
either a specific value for Ui (if v falls between two breakpoints of FUi (v)) , or a range
of allowed values for Ui (if v is equal to a breakpoint). Let us assume without loss of
generality that no two FUi (v) share a breakpoint, and let Vo = 0 < VI < ... < Vs = 1
be the ordered union of the breakpoints of the FUi (v). Thus for any breakpoint Vi,
there is at most one distinguished parent Uj (that we shall call the free parent) for
which Fu; (Vi) specifies an allowed interval of play for Uj . All other Ui are assigned
fixed values by Fu; (ve). For each breakpoint Ve, we now define the set of values for
the child W that, as we let the free parent range across its allowed interval, permit
V to play any mixed strategy as a best response.
< VI < ... < Vs = 1 be the ordered union of the breakpoints of the parent policies Fu; (v). Fix any breakpoint Ve, and assume without loss
of generality that U I is the free parent of V for V = Ve. Let [a, b] be the allowed
interval ofUI specified by FUI (ve), and letui = Fu;(ve) for all 2 :::; i:::; k. We define
Definition 3 Let Vo = 0
W e = {w E [0,1]: (:lUI E [a,b])6.v(UI,U2, ... ,Uk,W) = O}.
In words, We is the set of values that W can play that allow V to play any mixed
strategy, preserving the existence of an upstream Nash from V given W = w.
The next lemma, which we state without proof and is a special case of Lemma 6 in
Kearns et al. (2001), limits the complexity of the sets We. It also follows from the
earlier work that We can be computed in time proportional to the size of V's local
game matrix - O(2k) for a vertex with k parents.
We say that an interval [a, b]
~
[0, 1] is floating if both a -I- 0 and b -I- 1.
Lemma 4 For any breakpoint Ve, the set We is either empty, a single interval, or
the union of two intervals that are not floating.
We wish to create the (inductive) Nash breakpoint policy Fv(w) from the sets W e
and the Fu; policies. The idea is that if w E We for some breakpoint index e,
then by definition of W e, if W plays wand the Uis play according to the setting
determined by the Fu; policies (including a fixed setting for the free parent of V),
any play by V is a best response-so in particular, V may play the breakpoint value
Ve, and thus extend the Nash solution constructed, as the UiS can also all be best
responses. For b E {O, I}, we define W b as the set of values w such that if W = w
and the Uis are set according to their breakpoint policies for V = b, V = b is a
best response. To create Fv (w) as a total function, we must first show that every
w E [0, 1] is contained in some We or W O or WI.
Lemma 5 Let Vo = 0 < VI < ... < Vs = 1 be the ordered union of the breakpoints
of the Fu; (v) policies. Then for any value w E [0, 1], either w E w b for some
bE {O, I} , or there exists an index e such that wE W e.
Proof: Consider any fixed value of w, and for each open interval (vi> vj+d determined by adjacent breakpoints, label this interval by V 's best response (0 or
1) to W = wand 0 set according to the Fu; policies for this interval. If either
the leftmost interval [O ,vd is labeled with 0 or the rightmost interval [v s -I , I] is
labeled with 1, then w is included in W O or WI , respectively (V playing 0 or 1
is a best response to what the Uis will play in response to a 0 or 1). Otherwise,
since the labeling starts at 1 on the left and ends at 0 on the right, there must be
a breakpoint Ve such that V's best response changes over this breakpoint. Let Ui
be the free parent for this breakpoint. By continuity, there must be a value of Ui
in its allowed interval for which V is indifferent to playing 0 or 1, so w E W e. This
completes the proof of Lemma 5.
Armed with Lemmas 4 and 5, we can now describe the construction of Fv(w). Since
every w is contained in some W e (Lemma 5), and since every W e is the union of at
most two intervals (Lemma 4), we can uniquely identify the set WeI that covers the
largest (leftmost) interval containing w = 0; let [0, a] be this interval. Continuing
in the same manner to the right, we can identify the unique set We2 that contains
v7r - - - - - --- ----- --- ----- --- ----- --- ----- -------- - r - - - - - -
v6 -
V
------~
--------
v5 ------------------------------ -
v4 -
-
--
-
--
-
--
-
--
-
-----
------------ - --
, - - - - - - - ' ---------------
------------- ------------ ---------------------
v3f------.- ---------------------- --------------------------- - -
v2 _______--'-_ _ _ _ _ _----L _________________________________ _
vI ----------------------------------------------------
w
Figure 1: Example of the inductive construction of Fv(w). The dashed horizontal
lines show the vrbreakpoints determined by the parent policies Fu; (v). The solid
intervals along these breakpoints are the sets We. As shown in Lemma 4, each of
these sets consists of either a single (possibly floating) interval, or two non-floating
intervals. As shown in Lemma 5, each value of w is covered by some We. The
construction of Fv(w) (represented by a thick line) begins on the left, and always
next "jumps" to the interval allowing greatest progress to the right.
w = a and extends farthest to the right of a. Any overlap between We 1 and We 2
can be arbitrarily assigned coverage by We 1 , and We2 "trimmed" accordingly; see
Figure 1. This process results in a Nash breakpoint policy Fv(w).
Finally, we bound the number of breakpoints in the Fv (w) policy. By construction,
each of its breakpoints must be the rightmost portion of some interval in WO, WI, or
some We. After the first breakpoint, each of these sets contributes at most one new
breakpoint (Lemma 4). The final breakpoint is at w = 1 and does not contribute
to the count (Definition 1). There is at most one We for each breakpoint in each
Fu; (v) policy, plus WO and WI, plus the initial leftmost interval and minus the
final breakpoint, so the total breakpoints in Fv(w) can be no more than two plus
the total number of breakpoints in the Fu; (v) policies. Therefore, the root of a size
n tree will have a Nash breakpoint policy with no more than 2n breakpoints.
This completes the proof of Theorem 2.
3.2
Upstream Pass
The downstream pass completes when each vertex in the tree has had its Nash
breakpoint policy computed. For simplicity of description, imagine that the root of
t he tree includes a dummy child with constant payoffs and no influence on t he root,
so the root's breakpoint policy has t he same form as the others in the tree.
To produce a Nash equilibrium, our algorithm performs an upstream pass over
the tree, starting from the root. Each vertex is told by its child what value to
play, as well as the value the child itself will play. The algorithm ensures that all
downstream vertices are Nash (playing best response to their neighbors). Given
this information, each vertex computes a value for each of its parents so that its
own assigned action is a best response. This process can be initiated by the dummy
vertex picking an arbitrary value for itself, and selecting the root's value according
to its Nash breakpoint policy.
Inductively, we have a vertex V connected to parents U1 , ... , Uk (or no parents if
V is a leaf) and child W. The child of V has informed V to chose V = v and that
it will play W = w. To decide on values for V's parents to enforce V playing a best
response, we can look at the Nash breakpoint policies FUi (v), which provide a value
(or range of values) for Ui as a function of v that guarantee an upstream Nash. The
value v can be a breakpoint for at most one Ui . For each Ui , if v is not a breakpoint
in FUi (v) , then Ui should be told to select Ui = FUi (v). If v is a breakpoint in
FUi (v), then Ui's value can be computed by solving ~V(Ul "'" Ui,"" Uk, w) = 0;
this is the value of Ui that makes V indifferent. The equation is linear in Ui and has
a solution by the construction of the Nash breakpoint policies on the downstream
pass. Parents are passed their assigned values as well as the fact that V = v.
When the upstream pass completes, each vertex has a concrete choice of action such
that jointly they have formed a Nash equilibrium.
The total running time of the algorithm can be bounded as follows. Each vertex
is involved in a computation in the downstream pass and in the upstream pass.
Let t be the total number of breakpoints in the breakpoint policy for a vertex V
with k parents. Sorting the breakpoints and computing the W? sets and computing
the new breakpoint policy can be completed in 0 (t log t + t2 k ). In the upstream
pass, only one breakpoint is considered, so 0 (log t + 2k) is sufficient for passing
breakpoints to the parents. By Theorem 2, t :S 2n , so the entire algorithm executes
in time O(n 2 10g n + n22k), where k is the largest number of neighbors of any vertex
in the network.
The algorithm can be implemented to take advantage of local game matrices provided in a parameterized form. For example, if each vertex's payoff is solely a
function of the number of 1s played by the vertex's neighbors, the algorithm takes
O(n 2 10gn + n 2 k), eliminating the exponential dependence on k.
4
Conclusion
The algorithm presented in this paper finds a single Nash equilibrium for a game
represented by a tree-structured network. By building representations of all equilibria, our earlier algorithm (Kearns et al. 2001) was able to select equilibria efficiently
according to criteria like maximizing the total expected payoff for all players. The
polynomial-time algorithm described in this paper throws out potential equilibria
at many stages, most significantly during the construction of the Nash breakpoint
policies. An interesting area for future work is to manipulate this process to produce
equilibria with particular properties.
References
Michael Kearns, Michael L. Littman, and Satinder Singh. Graphical models for game
theory. In Proceedings of the 17th Conference on Uncertainty in Artificial Int elligence
(UAI), pages 253- 260, 200l.
Daphne Koller and Brian Milch. Multi-agent influence diagrams for representing and
solving games. Submitted, 2001.
Pierfrancesco La Mura. Game networks. In Proceedings of the 16th Conference on Uncertainty in Artificial Intelligence (UAI), pages 335- 342, 2000.
J . F. Nash. Non-cooperative games. Annals of Math ematics, 54:286- 295, 1951.
| 2101 |@word eliminating:1 polynomial:4 replicate:1 open:2 minus:1 solid:1 accommodate:1 initial:1 mkearns:1 att:1 contains:1 selecting:1 rightmost:2 com:1 assigning:1 must:4 gv:1 v:3 intelligence:1 leaf:5 accordingly:1 ith:1 math:1 contribute:1 daphne:1 along:1 constructed:1 consists:3 manner:1 pairwise:1 upenn:1 indeed:1 expected:4 behavior:1 multi:8 armed:1 becomes:1 provided:3 distri:1 underlying:2 notation:2 bounded:2 begin:1 what:2 informed:1 finding:1 nj:1 guarantee:1 every:5 ful:1 exactly:4 demonstrates:1 uk:5 farthest:1 before:1 positive:2 local:12 treat:1 limit:1 initiated:1 path:2 solely:1 abuse:1 plus:4 chose:1 range:4 directed:1 unique:1 union:5 area:1 significantly:1 word:2 suggest:1 milch:2 influence:2 collapsed:1 maximizing:1 starting:1 simplicity:1 pure:1 unilaterally:1 his:2 classic:1 analogous:1 annals:1 construction:6 play:22 colorado:1 suppose:1 exact:3 imagine:1 pa:1 trick:1 crossing:2 labeled:3 cooperative:1 ensures:1 connected:1 nash:41 complexity:2 ui:17 littman:2 inductively:1 singh:2 solving:3 joint:3 represented:5 describe:5 artificial:2 tell:1 labeling:1 choosing:1 neighborhood:2 say:5 otherwise:2 florham:1 commit:1 jointly:1 noisy:1 itself:3 final:2 advantage:1 interaction:3 product:2 syntek:1 representational:2 description:1 parent:21 empty:1 produce:2 leave:1 received:1 progress:1 solves:1 throw:1 coverage:1 c:1 implemented:1 convention:1 thick:1 fix:1 preliminary:1 brian:1 designate:1 extension:1 sufficiently:1 considered:1 equilibrium:26 label:1 largest:2 create:2 always:4 rather:1 progressing:1 inference:4 typically:1 entire:2 koller:2 i1:1 special:1 field:1 equal:2 saving:1 ng:2 represents:1 park:1 look:1 breakpoint:48 alter:1 tabular:3 future:2 others:1 t2:1 ve:9 deviating:1 familiar:1 floating:4 maintain:1 interest:2 indifferent:5 introduces:1 we2:2 ematics:1 edge:2 fu:11 partial:1 necessary:1 tree:15 continuing:1 desired:1 accrue:1 formalism:2 earlier:2 gn:1 cover:1 assignment:1 vertex:27 entry:4 subset:1 considerably:2 probabilistic:4 physic:1 v4:1 told:2 picking:1 fui:11 michael:4 analogously:1 concrete:1 von:1 satisfied:1 containing:1 choose:2 possibly:2 potential:1 includes:2 int:1 mv:4 vi:10 root:10 lab:1 portion:1 start:1 formed:1 accomplishing:1 efficiently:6 listing:1 v7r:1 identify:3 bayesian:3 executes:1 submitted:1 definition:7 involved:1 proof:6 mi:12 associated:1 recall:1 organized:1 appears:2 response:17 wei:1 generality:3 furthermore:1 just:1 stage:1 horizontal:1 continuity:1 quality:1 perhaps:1 building:1 inductive:4 assigned:5 adjacent:1 game:38 during:1 uniquely:1 criterion:1 leftmost:3 butions:1 theoretic:1 complete:1 vo:3 performs:1 extend:1 interpretation:2 slight:1 he:3 trivially:1 pm:1 baveja:1 had:1 base:2 mlittman:1 own:1 recent:2 arbitrarily:1 vt:1 seen:1 preserving:1 fuk:1 tempting:1 dashed:1 ing:1 cross:1 manipulate:1 himself:1 represent:1 want:1 addressed:1 interval:22 completes:4 diagram:1 macroscopic:1 w2:1 pass:1 undirected:2 integer:1 call:1 xj:1 pennsylvania:1 identified:1 idea:1 avenue:1 passed:3 trimmed:1 ul:1 wo:3 unilateral:1 york:1 passing:1 jj:2 action:8 useful:1 covered:1 involve:1 processed:1 simplest:1 specifies:3 sign:2 dummy:2 shall:7 incentive:1 express:1 nevertheless:1 capital:1 pj:2 uis:4 graph:5 symbolically:1 downstream:8 wand:2 parameterized:1 uncertainty:2 extends:1 throughout:1 decide:1 entirely:1 bound:2 breakpoints:20 played:2 ilk:2 u1:1 structured:3 department:1 according:5 smaller:1 mura:2 across:1 wi:8 invariant:1 equation:1 count:1 letting:1 end:1 permit:1 v2:1 enforce:1 distinguished:1 existence:1 original:1 denotes:1 running:1 include:1 completed:1 graphical:18 uj:2 classical:1 seeking:1 v5:1 realized:1 parametric:2 strategy:10 dependence:1 separating:1 ensuing:1 vd:1 extent:1 trivial:2 index:8 elligence:1 difficult:1 negative:2 policy:36 allowing:1 payoff:8 extended:3 arbitrary:1 introduced:3 pair:1 specified:2 fv:17 able:1 below:1 challenge:1 including:1 greatest:1 overlap:1 representing:2 improve:1 brief:1 axis:1 extract:1 philadelphia:1 deviate:1 understanding:1 marginalizing:1 loss:3 mixed:9 interesting:2 proportional:1 analogy:2 agent:11 sufficient:1 playing:4 pi:2 share:1 course:1 changed:1 free:5 allow:1 neighbor:8 fall:2 sparse:2 benefit:2 xn:1 computes:5 jump:1 compact:2 implicitly:1 satinder:2 global:4 uai:2 table:1 contributes:1 upstream:11 vj:1 main:2 nothing:1 child:13 allowed:5 centerpiece:1 ny:1 wish:1 exponential:4 xl:1 theorem:5 down:1 specific:1 exists:2 ci:1 subtree:1 sigmoids:1 sorting:1 easier:1 vii:1 intersection:1 simply:1 ordered:4 contained:2 v6:1 u2:1 conditional:2 towards:3 change:1 included:1 determined:5 except:2 lui:1 wt:2 kearns:7 lemma:11 called:4 total:7 pas:14 la:2 player:18 select:2 ex:1 |
1,210 | 2,102 | Products of Gaussians
Christopher K. I. Williams
Division of Informatics
University of Edinburgh
Edinburgh EH1 2QL, UK
c. k. i. [email protected]
http://anc.ed.ac.uk
Felix V. Agakov
System Engineering Research Group
Chair of Manufacturing Technology
Universitiit Erlangen-Niirnberg
91058 Erlangen, Germany
F.Agakov@lft?uni-erlangen.de
Stephen N. Felderhof
Division of Informatics
University of Edinburgh
Edinburgh EH1 2QL, UK
[email protected]
Abstract
Recently Hinton (1999) has introduced the Products of Experts
(PoE) model in which several individual probabilistic models for
data are combined to provide an overall model of the data. Below we consider PoE models in which each expert is a Gaussian.
Although the product of Gaussians is also a Gaussian, if each Gaussian has a simple structure the product can have a richer structure.
We examine (1) Products of Gaussian pancakes which give rise to
probabilistic Minor Components Analysis, (2) products of I-factor
PPCA models and (3) a products of experts construction for an
AR(l) process.
Recently Hinton (1999) has introduced the Products of Experts (PoE) model in
which several individual probabilistic models for data are combined to provide an
overall model of the data. In this paper we consider PoE models in which each
expert is a Gaussian. It is easy to see that in this case the product model will
also be Gaussian. However, if each Gaussian has a simple structure, the product
can have a richer structure. Using Gaussian experts is attractive as it permits a
thorough analysis of the product architecture, which can be difficult with other
models , e.g. models defined over discrete random variables.
Below we examine three cases of the products of Gaussians construction: (1) Products of Gaussian pancakes (PoGP) which give rise to probabilistic Minor Components Analysis (MCA), providing a complementary result to probabilistic Principal
Components Analysis (PPCA) obtained by Tipping and Bishop (1999); (2) Products of I-factor PPCA models; (3) A products of experts construction for an AR(l)
process.
Products of Gaussians
If each expert is a Gaussian pi(xI8 i ) '" N(J1i' ( i), the resulting distribution of the
product of m Gaussians may be expressed as
By completing the square in the exponent it may be easily shown that p(xI8)
N(/1;E, (2:), where (E l = 2::1 (i l . To simplify the following derivations we will
assume that pi(xI8 i ) '" N(O, (i) and thus that p(xI8) '" N(O, (2:). J12: i
can be
obtained by translation of the coordinate system.
?
1
Products of Gaussian Pancakes
A Gaussian "pancake" (GP) is a d-dimensional Gaussian, contracted in one dimension and elongated in the other d - 1 dimensions. In this section we show that the
maximum likelihood solution for a product of Gaussian pancakes (PoGP) yields a
probabilistic formulation of Minor Components Analysis (MCA).
1.1
Covariance Structure of a GP Expert
Consider a d-dimensional Gaussian whose probability contours are contracted
in the direction w and equally elongated in mutually orthogonal directions
VI , ... , vd-l.We call this a Gaussian pancake or GP. Its inverse covariance may be
written as
d- l
( -1=
L ViV; /30 + ww
T
/3,;;,
(1)
i= l
where VI, ... ,V d - l, W form a d x d matrix of normalized eigenvectors of the covariance C. /30 = 0"0 2 , /3,;; = 0";;2 define inverse variances in the directions of elongation
and contraction respectively, so that 0"5 2 0"1. Expression (1) can be re-written in
a more compact form as
(2)
where w = wJ/3,;; - /30 and Id C jRdxd is the identity matrix. Notice that according
to the constraint considerations /30 < /3,;;, and all elements of ware real-valued.
Note the similarity of (2) with expression for the covariance of the data of a 1factor probabilistic principal component analysis model ( = 0"21d + ww T (Tipping
and Bishop, 1999) , where 0"2 is the variance of the factor-independent spherical
Gaussian noise. The only difference is that it is the inverse covariance matrix for
the constrained Gaussian model rather than the covariance matrix which has the
structure of a rank-1 update to a multiple of Id .
1.2
Covariance of the PoGP Model
We now consider a product of m GP experts, each of which is contracted in a single
dimension. We will refer to the model as a (I,m) PoGP, where 1 represents the
number of directions of contraction of each expert. We also assume that all experts
have identical means.
From (1), the inverse covariance of the the resulting (I,m) PoGP model can be
expressed as
m
C;;l
=L
Ci l
(3)
i=l
where columns of We Rdxm correspond to weight vectors of the m PoGP experts,
and (3E = 2::1 (3~i) > o.
1.3
Maximum-Likelihood Solution for PoGP
Comparing (3) with m-factor PPCA we can make a conjecture that in contrast
with the PPCA model where ML weights correspond to principal components of
the data covariance (Tipping and Bishop, 1999), weights W of the PoGP model
define projection onto m minor eigenvectors of the sample covariance in the visible
d-dimensional space, while the distortion term (3E Id explains larger variations l . This
is indeed the case.
In Williams and Agakov (2001) it is shown that stationarity of the log-likelihood
with respect to the weight matrix Wand the noise parameter (3E results in three
classes of solutions for the experts' weight matrix, namely
W
5
5W
0;
CE ;
CEW, W:j:. 0, 5:j:. CE ,
(4)
where 5 is the covariance matrix of the data (with an assumed mean of zero). The
first two conditions in (4) are the same as in Tipping and Bishop (1999), but for
PPCA the third condition is replaced by C-l W = 5- l W (assuming that 5 - 1 exists).
In Appendix A and Williams and Agakov (2001) it is shown that the maximum
likelihood solution for WML is given by:
(5)
where R c Rmxm is an arbitrary rotation matrix, A is a m x m matrix containing
the m smallest eigenvalues of 5 and U = [Ul , ... ,u m ] c Rdxm is a matrix of the
corresponding eigenvectors of 5. Thus, the maximum likelihood solution for the
weights of the (1, m) PoG P model corresponds to m scaled and rotated minor
eigenvectors of the sample covariance 5 and leads to a probabilistic model of minor
component analysis. As in the PPCA model, the number of experts m is assumed
to be lower than the dimension of the data space d.
The correctness of this derivation has been confirmed experimentally by using a
scaled conjugate gradient search to optimize the log likelihood as a function of W
and (3E.
1.4
Discussion of PoGP model
An intuitive interpretation of the PoGP model is as follows: Each Gaussian pancake
imposes an approximate linear constraint in x space. Such a linear constraint is that
x should lie close to a particular hyperplane. The conjunction of these constraints
is given by the product of the Gaussian pancakes. If m ? d it will make sense to
lBecause equation 3 has the form of a factor analysis decomposition, but for the inverse
covariance matrix, we sometimes refer to PoGP as the rotcaf model.
define the resulting Gaussian distribution in terms of the constraints. However, if
there are many constraints (m > d/2) then it can be more efficient to describe the
directions of large variability using a PPCA model, rather than the directions of
small variability using a PoGP model. This issue is discussed by Xu et al. (1991) in
what they call the "Dual Subspace Pattern Recognition Method" where both PCA
and MCA models are used (although their work does not use explicit probabilistic
models such as PPCA and PoGP).
MCA can be used , for example, for signal extraction in digital signal processing
(Oja, 1992), dimensionality reduction, and data visualization. Extraction of the
minor component is also used in the Pisarenko Harmonic Decomposition method
for detecting sinusoids in white noise (see, e.g. Proakis and Manolakis (1992), p .
911). Formulating minor component analysis as a probabilistic model simplifies
comparison of the technique with other dimensionality reduction procedures, permits extending MCA to a mixture of MCA models (which will be modeled as a
mixture of products of Gaussian pancakes) , permits using PoGP in classification
tasks (if each PoGP model defines a class-conditional density) , and leads to a number of other advantages over non-probabilistic MCA models (see the discussion of
advantages of PPCA over PCA in Tipping and Bishop (1999)).
2
Products of PPCA
In this section we analyze a product of m I-factor PPCA models, and compare it
to am-factor PPCA model.
2.1
I-factor PPCA model
Consider a I-factor PPCA model, having a latent variable Si and visible variables x.
The joint distribution is given by P(Si, x) = P(si) P(xlsi). We set P(Si) '" N(O, 1)
and P(XI Si) '" N(WiSi' (]"2) . Integrating out Si we find that Pi(x) '" N(O, Ci ) where
C = wiwT + (]"21d and
(6)
where (3 = (]"-2 and "(i = (3/(1 + (3 llwi W). (3 and "(i are the inverse variances in the
directions of contraction and elongation respectively.
The joint distribution of Si and x is given by
(7)
[s;
exp - -(3 - - 2x T WiSi
2 "(i
+ XT X]
.
(8)
Tipping and Bishop (1999) showed that the general m-factor PPCA model (mPPCA) has covariance C = (]"21d + WW T , where W is the d x m matrix of factor
loadings. When fitting this model to data, the maximum likelihood solution is to
choose W proportional to the principal components of the data covariance matrix.
2.2
Products of I-factor PPCA models
We now consider the product of m I-factor PPCA models, which we denote a
(1, m)-PoPPCA model. The joint distribution over 5 = (Sl' ... ,Srn)T and x is
13
P(x,s) ex: exp -"2
L [s;- -:- - 2x
m
i=l
T
W iSi
+ XT X ]
(9)
?
,,(,
Let zT d~f (xT , ST). Thus we see that the distribution of z is Gaussian with inverse
covariance matrix 13M, where
-W)
r -1
(10)
,
and r = diag("(l , ... ,"(m)' Using the inversion equations for partitioned matrices
(Press et al., 1992, p. 77) we can show that
(11)
where ~xx is the covariance of the x variables under this model. It is easy to confirm
that this is also the result obtained from summing (6) over i = 1, ... ,m.
2.3
Maximum Likelihood solution for PoPPCA
Am-factor PPCA model has covariance a21d + WW T and thus, by the Woodbury
formula, it has inverse covariance j3 ld - j3W(a2 lm + WT W) - lW T . The maximum
likelihood solution for a m-PPCA model is similar to (5), i.e. W = U(A _a2Im)1/2 RT,
but now A is a diagonal matrix of the m principal eigenvalues, and U is a matrix
of the corresponding eigenvectors. If we choose RT = I then the columns of W are
orthogonal and the inverse covariance of the maximum likelihood m-PPCA model
has the form j3 ld - j3WrwT. Comparing this to (11) (with W = W) we see that the
difference is that the first term of the RHS of (11) is j3m1d , while for m-PPCA it is
j3 ld.
In section 3.4 and Appendix C.3 of Agakov (2000) it is shown that (for m
obtain the m-factor PPCA solution when
-
m
-
A<A' < - - A
- '
m -I '
i = 1, ...
,m,
:::=:
2) we
(12)
where Ais the mean of the d - m discarded eigenvalues, and Ai is a retained eigenvalue; it is the smaller eigenvalues that are discarded. We see that the covariance
must be nearly spherical for this condition to hold. For covariance matrices satisfying (12) , this solution was confirmed by numerical experiments as detailed in
(Agakov, 2000, section 3.5).
To see why this is true intuitively, observe that Ci 1 for each I-factor PPCA expert
will be large (with value 13) in all directions except one. If the directions of contraction for each Ci 1 are orthogonal, we see that the sum of the inverse covariances
will be at least (m - 1)13 in a contracted direction and m j3 in a direction in which
no contraction occurs. The above shows that for certain types of sample covariance matrix the (1 , m) PoPPCA solution is not equivalent to the m-factor PPCA
solution. However, it is interesting to note that by relaxing the constraint on the
isotropy of each expert's noise the product of m one-factor factor analysis models
can be shown to be equivalent to an m-factor factor analyser (Marks and Movellan,
2001).
???
?? ?
?
(c)
(b)
?
(d)
Figure 1: (a) Two experts. The upper one depicts 8 filled circles (visible units) and
4 latent variables (open circles), with connectivity as shown. The lower expert also
has 8 visible and 4 latent variables, but shifted by one unit (with wraparound). (b)
Covariance matrix for a single expert. (c) Inverse covariance matrix for a single
expert. (d) Inverse covariance for product of experts.
3
A Product of Experts Representation for an AR(l)
Process
For the PoPPCA case above we have considered models where the latent variables
have unrestricted connectivity to the visible variables. We now consider a product
of experts model with two experts as shown in Figure l(a). The upper figure depicts
8 filled circles (visible units) and 4 latent variables (open circles), with connectivity
as shown. The lower expert also has 8 visible and 4 latent variables, but shifted
by one unit (with wraparound) with respect to the first expert. The 8 units are,
of course, only for illustration- the construction is valid for any even number of
visible units.
Consider one hidden unit and its two visible children. Denote the hidden unit by s
the visible units as Xl and Xr (l, r for left and right). Set s '" N(O, 1) and
Xl
= as
+ bWI
Xr
= ?as + bwr
(13)
,
where WI and Wr are independent N(O , 1) random variables, and a , b are constants.
(This is a simple example of a Gaussian tree-structured process, as studied by a
number of groups including that led by Prof. Willsky at MIT; see e.g. Luettgen
et al. (1993).) Then (xf) = (x;) = a2 + b2 and (XIX r ) = ?a2 ? The corresponding
2 x 2 inverse covariance matrix has diagonal entries of (a 2 + b2 )j ~ and off-diagonal
entries of =t=a 2 j~ , where ~ = b2 (b 2 + 2a 2 ).
Graphically, the covariance matrix of a single expert has the form shown in Figure
l(b) (where we have used the + rather than - choice from (13) for all variables).
Figure l(c) shows the corresponding inverse covariance for the single expert, and
Figure 1(d) shows the resulting inverse covariance for the product of the two experts,
with diagonal elements 2(a 2 + b2 )j ~ and off-diagonal entries of =t=a 2 j~.
An AR(l) process of the circle with d nodes has the form Xi = aXi -
1 (mod d)
+ Zi,
where Zi ~ N(O,v). Thusp(X) <X exp-21v L:i(Xi-aXi - 1 (mod d))2 and the inverse
covariance matrix has a circulant tridiagonal structure with diagonal entries of
(1 + ( 2 )/v and off-diagonal entries of -a/v. The product of experts model defined
above can be made equivalent to the circular AR(I) process by setting
(14)
The ? is needed in (13) as when a is negative we require
the inverse covariances.
Xr
= -as + bWr
to match
We have shown that there is an exact construction to represent a stationary circular AR(I) process as a product of two Gaussian experts. The approximation
of other Gaussian processes by products of tree-structured Gaussian processes is
further studied in (Williams and Felderhof, 2001). Such constructions are interesting because they may allow fast approximate inference in the case that d is large
(and the target process may be 2 or higher dimensional) and exact inference is not
tractable. Such methods have been developed by Willsky and coauthors, but not
for products of Gaussians constructions.
Acknowledgements
This work is partially supported by EPSRC grant GR/L78161 Probabilistic Models
for Sequences. Much of the work on PoGP was carried out as part of the MSc
project of FVA at the Division of Informatics, University of Edinburgh. CW thanks
Sam Roweis, Geoff Hinton and Zoubin Ghahramani for helpful conversations on the
rotcaf model during visits to the Gatsby Computational Neuroscience Unit. FVA
gratefully acknowledges the support of the Royal Dutch Shell Group of Companies
for his MSc studies in Edinburgh through a Centenary Scholarship. SNF gratefully
acknowledges additional support from BAE Systems.
References
Agakov, F. (2000). Investigations of Gaussian Products-of-Experts Models. Master's
thesis, Division of Informatics, The University of Edinburgh. Available at http://'iI'iI'iI .
dai.ed.ac.uk/homes/felixa/all.ps.gz.
Hinton, G . E . (1999) . Products of experts. In Proceedings of the Ninth International
Conference on Artificial N eural N etworks (ICANN gg), pages 1- 6.
Luettgen , M. , Karl, W. , and Willsky, A. (1993). Multiscale Representations of Markov
Random Fields. IEEE Trans. Signal Processing, 41(12):3377- 3395.
Marks , T. and Movellan, J. (2001). Diffusion Networks, Products of Experts, and Factor
Analysis. In Proceedings of the 3rd International Conference on Independent Component
Analysis and Blind Source Separation.
OJ a, E. (1992). Principal Components, Minor Components, and Linear Neural Networks.
Neural N etworks, 5:927 - 935.
Press, W. H. , Teukolsky, S. A., Vetterling, W. T., and Flannery, B. P. (1992). Num erical
Recipes in C. Cambridge University Press, Second edition .
Proakis, J. G. and Manolakis, D. G. (1992). Digital Signal Processing: Principles, Algorithms and Applications. Macmillan.
Tipping, M. E. and Bishop, C. M. (1999). Probabilistic principal components analysis. J.
Roy. Statistical Society B, 61(3) :611- 622.
Williams, C. K. I. and Agakov, F. V. (2001). Products of Gaussians and Probabilistic
Minor Components Analysis. Technical Report EDI-INF-RR-0043 , Division of Informatics, University of Edinburgh. Available at http://'iI'iI'iI. informatics. ed. ac. ukl
publications/report/0043.html.
Williams, C. K. I. and Felderhof, S. N. (2001). Products and Sums of Tree-Structured
Gaussian Processes. In Proceedings of the ICSC Symposium on Soft Computing 2001
(SOCO 2001).
Xu, L. and Krzyzak, A. and Oja, E. (1991). Neural Nets for Dual Subspace Pattern
Recogntion Method. International Journal of Neural Systems, 2(3):169- 184.
A
ML Solutions for PoGP
Here we analyze the three classes of solutions for the model covariance matrix which
result from equation (4) of section 1.3.
The first case W
= 0 corresponds to a
minimum of the log-likelihood.
In the second case, the model covariance e~ is equal to the sample covariance 5.
From expression (3) for e i;l we find WWT = 5 - 1 - ;3~ ld. This has the known
solution W = Um(A - 1 - ;3~ lm)1 /2 RT , where Um is the matrix of the m eigenvectors
of 5 with the smallest eigenvalues and A is the corresponding diagonal matrix of the
eigenvalues. The sample covariance must be such that the largest d - m eigenvalues
are all equal to ;3~; the other m eigenvalues are matched explicitly.
Finally, for the case of approximate model covariance (5W = e~w, 5 =f. e~) we, by
analogy with Tipping and Bishop (1999), consider the singular value decomposition
of the weight matrix, and establish dependencies between left singular vectors of
W = ULR T and eigenvectors of the sample covariance 5. U = [U1 , U2 , ... , um] C
lRdxm is a matrix of left singular vectors of W with columns constituting an orthonormal basis, L = diag(h,l2, ... ,lm) C lRmxm is a diagonal matrix of the singular values of Wand R C lRmxm defines an arbitrary rigid rotation of W. For this
case equation (4) can be written as 5UL = e ~ UL , where e~ is obtained from (3) by
applying the matrix inversion lemma [see e.g. Press et al. (1992)]. This leads to
5UL = e~UL
(;3i;lld - ;3i;l W(;3~ + WTW) -l WT )UL
U(;3i;l lm - ;3i;l LRT(;3~ lm + RL2RT) -l RL)L
U(;3i; l lm - ;3i;l(;3~ L -2 + Im) -l) L.
(15)
Notice that the term ;3;1 1m - ;3;l(;3~ L -2 + Im)-l in the r.h.s. of equation (15) is
just a scaling factor of U. Equation (15) defines the matrix form of the eigenvector
equation, with both sides post-multiplied by the diagonal matrix L.
If li
=f. 0 then (15) implies that
e ~ U i = 5Ui = AiUi, Ai = ;3i;1(1 - (;3~li2 + 1) - 1),
(16)
where Ui is an eigenvector of 5, and Ai is its corresponding eigenvalue. The scaling
factor li of the ith retained expert can be expressed as li = (Ail - ;3~)1/2 .
Obviously, if li = 0 then Ui is arbitrary. If li = 0 we say that the direction corresponding to Ui is discarded, i.e. the variance in that direction is explained merely
by noise. Otherwise we say that Ui is retained. All potential solutions of W may
then be expressed as
W = Um(D - ;3~ lm)1 /2 RT ,
(17)
where R C lRmxm is a rotation matrix, Um = [U1U2 ... um] C lRdxm is a matrix whose
columns correspond to m eigenvectors of 5, and D = diag( d 1 , d 2 , ... , dm ) C lRm xm
such that di = Ail if Ui is retained and d i = ;3~ if Ui is discarded.
It may further be shown (Williams and Agakov (2001)) that the optimal solution
for the likelihood is reached when W corresponds to the minor eigenvectors of the
sample covariance 5.
| 2102 |@word inversion:2 loading:1 open:2 contraction:5 covariance:40 decomposition:3 ld:4 reduction:2 comparing:2 si:7 written:3 must:2 numerical:1 visible:10 update:1 stationary:1 ith:1 num:1 detecting:1 node:1 symposium:1 fitting:1 indeed:1 isi:1 examine:2 lrmxm:3 wml:1 spherical:2 company:1 project:1 xx:1 matched:1 isotropy:1 what:1 ail:2 eigenvector:2 developed:1 lft:1 thorough:1 um:6 scaled:2 uk:6 unit:10 grant:1 felix:1 engineering:1 id:3 ware:1 lrm:1 studied:2 relaxing:1 woodbury:1 movellan:2 xr:3 procedure:1 snf:1 projection:1 integrating:1 zoubin:1 onto:1 close:1 applying:1 optimize:1 equivalent:3 elongated:2 williams:8 graphically:1 felderhof:3 orthonormal:1 his:1 j12:1 coordinate:1 variation:1 construction:7 target:1 exact:2 element:2 roy:1 recognition:1 satisfying:1 agakov:9 epsrc:1 wj:1 ui:7 division:5 basis:1 easily:1 joint:3 geoff:1 derivation:2 fast:1 describe:1 artificial:1 analyser:1 whose:2 richer:2 larger:1 valued:1 mppca:1 distortion:1 say:2 otherwise:1 gp:4 obviously:1 advantage:2 eigenvalue:10 sequence:1 rr:1 net:1 xi8:4 product:40 roweis:1 intuitive:1 recipe:1 p:1 extending:1 rotated:1 ac:5 minor:11 implies:1 direction:13 explains:1 require:1 investigation:1 im:2 hold:1 considered:1 exp:3 lm:7 smallest:2 a2:3 largest:1 correctness:1 mit:1 gaussian:29 rather:3 publication:1 conjunction:1 rank:1 likelihood:12 contrast:1 sense:1 am:2 helpful:1 inference:2 rigid:1 vetterling:1 hidden:2 germany:1 lrt:1 overall:2 issue:1 dual:2 classification:1 proakis:2 exponent:1 html:1 constrained:1 field:1 equal:2 ukl:1 extraction:2 having:1 elongation:2 identical:1 represents:1 wisi:2 nearly:1 report:2 simplify:1 oja:2 individual:2 replaced:1 recogntion:1 stationarity:1 circular:2 mixture:2 orthogonal:3 pancake:9 filled:2 tree:3 srn:1 re:1 circle:5 erical:1 column:4 soft:1 ar:6 entry:5 tridiagonal:1 gr:1 dependency:1 combined:2 st:1 density:1 thanks:1 international:3 probabilistic:14 off:3 informatics:6 connectivity:3 thesis:1 luettgen:2 containing:1 choose:2 expert:36 li:5 potential:1 de:1 b2:4 felixa:1 explicitly:1 vi:2 blind:1 icsc:1 analyze:2 reached:1 aiui:1 square:1 variance:4 yield:1 correspond:3 confirmed:2 ed:5 dm:1 di:1 erlangen:3 ppca:25 conversation:1 dimensionality:2 wwt:1 higher:1 tipping:8 formulation:1 just:1 msc:2 christopher:1 multiscale:1 defines:3 normalized:1 true:1 sinusoid:1 white:1 attractive:1 llwi:1 during:1 gg:1 harmonic:1 consideration:1 recently:2 rotation:3 rl:1 bwi:1 discussed:1 interpretation:1 refer:2 cambridge:1 ai:4 rd:1 gratefully:2 similarity:1 showed:1 inf:1 certain:1 minimum:1 dai:2 mca:7 unrestricted:1 additional:1 signal:4 stephen:1 ii:6 multiple:1 technical:1 xf:1 match:1 post:1 equally:1 visit:1 j3:4 coauthor:1 dutch:1 sometimes:1 represent:1 singular:4 source:1 mod:2 call:2 easy:2 manolakis:2 zi:2 architecture:1 simplifies:1 poe:4 expression:3 pca:2 ul:6 krzyzak:1 detailed:1 eigenvectors:9 http:3 sl:1 notice:2 shifted:2 neuroscience:1 wr:1 li2:1 discrete:1 group:3 ce:2 wtw:1 diffusion:1 bae:1 merely:1 sum:2 wand:2 inverse:17 master:1 separation:1 home:1 appendix:2 scaling:2 completing:1 constraint:7 u1:1 chair:1 formulating:1 conjecture:1 structured:3 according:1 conjugate:1 smaller:1 lld:1 sam:1 partitioned:1 wi:1 intuitively:1 explained:1 equation:7 visualization:1 mutually:1 etworks:2 needed:1 tractable:1 available:2 gaussians:7 permit:3 multiplied:1 observe:1 rmxm:1 ghahramani:1 prof:1 scholarship:1 establish:1 society:1 eh1:2 occurs:1 rt:4 diagonal:10 gradient:1 subspace:2 cw:1 vd:1 willsky:3 assuming:1 modeled:1 retained:4 illustration:1 providing:1 ql:2 difficult:1 negative:1 rise:2 zt:1 upper:2 markov:1 discarded:4 hinton:4 variability:2 ww:4 ninth:1 arbitrary:3 wraparound:2 introduced:2 edi:1 namely:1 trans:1 below:2 pattern:2 xm:1 including:1 royal:1 oj:1 technology:1 carried:1 acknowledges:2 gz:1 acknowledgement:1 l2:1 interesting:2 proportional:1 analogy:1 digital:2 imposes:1 principle:1 pi:3 translation:1 karl:1 course:1 supported:1 side:1 allow:1 circulant:1 edinburgh:8 xix:1 dimension:4 axi:2 valid:1 contour:1 made:1 viv:1 constituting:1 approximate:3 compact:1 uni:1 confirm:1 ml:2 summing:1 assumed:2 xi:3 search:1 latent:6 why:1 anc:1 diag:3 icann:1 rh:1 noise:5 edition:1 child:1 complementary:1 xu:2 eural:1 contracted:4 depicts:2 gatsby:1 explicit:1 xl:2 lie:1 lw:1 third:1 formula:1 bishop:8 xt:3 exists:1 ci:4 flannery:1 led:1 expressed:4 macmillan:1 partially:1 u2:1 corresponds:3 teukolsky:1 shell:1 conditional:1 identity:1 manufacturing:1 pog:1 universitiit:1 experimentally:1 except:1 hyperplane:1 wt:2 principal:7 lemma:1 mark:2 support:2 ex:1 |
1,211 | 2,103 | ACh, Uncertainty, and Cortical Inference
Peter Dayan
Angela Yu
Gatsby Computational Neuroscience Unit
17 Queen Square, London, England, WC1N 3AR.
[email protected]
[email protected]
Abstract
Acetylcholine (ACh) has been implicated in a wide variety of
tasks involving attentional processes and plasticity. Following
extensive animal studies, it has previously been suggested that
ACh reports on uncertainty and controls hippocampal, cortical and
cortico-amygdalar plasticity. We extend this view and consider
its effects on cortical representational inference, arguing that ACh
controls the balance between bottom-up inference, influenced by
input stimuli, and top-down inference, influenced by contextual
information. We illustrate our proposal using a hierarchical hidden Markov model.
1 Introduction
The individual and joint computational roles of neuromodulators such as
dopamine, serotonin, norepinephrine and acetylcholine are currently the focus of
intensive study.5, 7, 9?11, 16, 27 A rich understanding of the effects of neuromodulators
on the dynamics of networks has come about through work in invertebrate systems.21 Further, some general computational ideas have been advanced, such as
that they change the signal to noise ratios of cells. However, more recent studies,
particularly those focusing on dopamine,26 have concentrated on specific computational tasks.
ACh was one of the first neuromodulators to be attributed a specific role. Hasselmo and colleagues,10, 11 in their seminal work, proposed that cholinergic (and,
in their later work, also GABAergic12 ) modulation controls read-in to and readout from recurrent, attractor-like memories, such as area CA3 of the hippocampus.
Such memories fail in a characteristic manner if the recurrent connections are operational during storage, thus forcing new input patterns to be mapped to existing
memories. Not only would these new patterns lose their specific identity, but,
worse, through standard synaptic plasticity, the size of the basin of attraction of
the offending memory would actually be increased, making similar problems more
likely. Hasselmo et al thus suggested, and collected theoretical and experimental
evidence in favor of, the notion that ACh (from the septum) should control the suppression and plasticity of specific sets of inputs to CA3 neurons. During read-in,
high levels of ACh would suppress the recurrent synapses, but make them readily
plastic, so that new memories would be stored without being pattern-completed.
Then, during read-out, low levels of ACh would boost the impact of the recurrent
weights (and reduce their plasticity), allowing auto-association to occur.
The ACh signal to the hippocampus can be characterized as reporting the unfamiliarity of the input with which its release is associated. This is analogous to its
characterization as reporting the uncertainty associated with predictions in theories
of attentional influences over learning in classical conditioning.4 In an extensive
series of investigations in rats, Holland and his colleagues14, 15 have shown that
a cholinergic projection from the nucleus basalis to the (parietal) cortex is important when animals have to devote more learning (which, in conditioning, is essentially synonymous with paying incremental attention) to stimuli about whose
consequences the animal is uncertain.20 We have4 interpreted this in the statistical terms of a Kalman filter, arguing that the ACh signal reported this uncertainty,
thus changing plasticity appropriately. Note, however, that unlike the case of the
hippocampus, the mechanism of action of ACh in conditioning is not well understood.
In this paper, we take the idea that ACh reports on uncertainty one step farther.
There is a wealth of analysis-by-synthesis unsupervised learning models of cortical processing.1, 3, 8, 13, 17, 19, 23 In these, top-down connections instantiate a generative
model of sensory input; and bottom-up connections instantiate a recognition model,
which is the statistical inverse of the generative model, and maps inputs into categories established in the generative model. These models, at least in principle, permit stimuli to be processed according both to bottom-up input and top-down expectations, the latter being formed based on temporal context or information from
other modalities. Top-down expectations can resolve bottom-up ambiguities, permitting better processing. However, in the face of contextual uncertainty, top-down
information is useless. We propose that ACh reports on top-down uncertainty,
and, as in the case of area CA3, differentially modulates the strength of synaptic
connections: comparatively weakening those associated with the top-down generative model, and enhancing those associated with bottom-up, stimulus-bound
information.2 Note that this interpretation is broadly consistent with existent electrophysiology data, and documented effects on stimulus processing of drugs that
either enhance (eg cholinesterase inhibitors) or suppress (eg scopolamine) the action of ACh.6, 25, 28
There is one further wrinkle. In exact bottom-up, top-down, inference using a
generative model, top-down contextual uncertainty does not play a simple role.
Rather, all possible contexts are treated simultaneously according to the individual posterior probabilities that they currently pertain. Given the neurobiologically
likely scenario in which one set of units has to be used to represent all possible
contexts, this exact inferential solution is not possible. Rather, we propose that a
single context is represented in the activities of high level (presumably pre-frontal)
cortical units, and uncertainty associated with this context is represented by ACh.
This cholinergic signal then controls the balance between bottom-up and top-down
influences over inference.
In the next section, we describe the simple hierarchical generative model that we
use to illustrate our proposal. The ACh-based recognition model is introduced in
section 3 and discussed in section 4.
2 Generative and Recognition Models
Figure 1A shows a very simple case of a hierarchical generative model. The generative model is a form of hidden Markov model (HMM), with a discrete hidden
state , which will capture the idea of a persistent temporal context, and a twodimensional, real-valued, output . Crucially, there is an extra layer, between
and . The state is stochastically determined from , and controls which of a
set of 2d Gaussians (centered at the corners of the unit square) is used to generate
. In this austere case, is the model?s representation of , and the key inference
!
4
3
2
1
4
3
2
1
0
2
2
1
1
0
?1 ?1
0
200
400
2
1
0
?1
?1 0 1 2
Figure 1: Generative model. A) Three-layer model "$#&%('*),+.-/102#3%4'*)5+4-6/178#,9;: with
dynamics (< ) in the " layer ( =8> "?A@B"C?ED[email protected] K(L ), a probabilistic mapping ( M ) from ";NO0
( =8> 0(?@P"C?RQ "[email protected] LV ), and a Gaussian model WX> 7YQ 04G with means at the corners of the unit
square and standard deviation IZJ V in each direction. The model is rotationally invariant;
only some of the links are shown for convenience. B) Sample sequence showing the slow
dynamics in " ; the stochastic mapping into 0 and the substantial overlap in 7 (different
symbols show samples from the different Gaussians shown in A).
problem will be to determine the distribution over which generated , given
the past experience [ ]\^`_ba ^4cddd c ]\ ^fe and itself.
Figure 1B shows an example of a sequence of gihih steps generated from the model.
The state in the layer stays the same for an average of about jZh timesteps; and
then switches to one of the other states, chosen equally at random. The transition
matrix is kTlnmSoqpRlRm . The state in the layer is more weakly determined by the state
in the layer, with a probability of only j*r(g that !_ . The stochastic transition
from to is governed by the transition matrix s l mHtum . Finally, is generated as a
Gaussian about a mean specified by . The standard deviation of these Gaussians
( h dv in each direction) is sufficiently large that the densities overlap substantially.
The naive solution to inferring is to use only the likelihood term (ie only the
probabilities wyx Cz H{ ). The performance of this is likely to be poor, since the Gaussians in for the different values of overlap so much. However, and this is why
it is a paradigmatic case for our proposal, contextual information, in this case past
experience, can help to determine . We show how the putative effect of ACh in
controlling the balance between bottom-up and top-down inference in this model
can be used to build a good approximate inference model.
In order to evaluate our approximate model, we need to understand optimal inference in this case. Figure 2A shows the standard HMM inference model, which
calculates the exact posteriors wyx Cz [ H{ and wyx Cz [ H{ . This is equivalent to just the
forward part of the forwards-backwards algorithm22 (since we are not presently
interested in learning the parameters of the model). The adaptation to include the
layer is straightforward. Figures 3A;D;E show various aspects of exact inference
for a particular run. The histograms in figure 3A show that wyx z [ { captures quite
well the actual states ! | that generated the data. The upper plot shows the posterior probabilities of the actual states in the sequence ? these should be, and are,
usually high; the lower histogram the posterior probability of the other possible
states; these should be, and are, usually low. Figure 3D shows the actual state sequence | ; figure 3E shows the states that are individually most likely at each time
step (note that this is not the maximum likelihood state sequence, as found by the
Viterbi algorithm, for instance).
(
)
*
% $
&
'
"!#
$
Figure 2: Recognition models. A) Exact recognition model. =8> " ?EDF Q + ?EDF G is propagated to
provide the prior =8> "? Q +`?ED F G (shown by the lengths of the thick vertical bars) and thus the
prior =8> 0(? Q +`?ED F G . This is combined with the likelihood term from the data 7? to give the true
=8> 0 ? Q + ? G . B) Bottom-recognition model uses only a generic prior over 0 ? (which conveys no
information), and so the likelihood term dominates. C) ACh model. A single estimated
state ", ?ED F is used, in conjunction with its certainty -6? DF , reported by cholinergic activity, to
, >."C, ?nQ/"C, ?ED F]G over "C? (which is a mixture of a delta function and
produce an approximate prior =8
a uniform), and thus an approximate prior over 0 ? . This is combined with the likelihood to
, > 0f? Q +`? G , and a new cholinergic signal -? is calculated.
give an approximate =8
V
021436587 9;:
1X
0.25
0 0
0.8
00
4 W
> 3
2
1
0
1Y
R
021=3 < 5 7 9;:
?@ABC?ED=> 5
A'\]H_^
R
PQ
1 NO
M
PQ
NM O
0
0
1
4Z
3
2
1
400 0
0UTE13S7 9K:
0 1 > 7 9K:
?EF&G
HI?J 2
A \]HI^
1
0
0
4[
3
2
1
400 0
02L 13S7 9K:
>L
A'\]H_^
1
400
Figure 3: Exact and approximation recognition. A) Histograms of the exact posterior distribution =8> 0!Q +`G over the actual state 0a? ` (upper) and the other possible states 0K@ b 0c? ` (lower,
written =8>'0 d ` G ). This shows the quality of exact representational inference. B;C) Comparison
, > 0(?nQ + G (C) with
of the purely bottom up =e4> 0f? Q 7?SG (B) and the ACh-based approximation =8
the true =8> 0(? Q +`G across all values of 0 . The ACh-based approximation is substantially more
accurate. D) Actual " ? . E) Highest probability " state from the exact posterior distribution.
F) Single " , state in the ACh model.
Figure 2B shows a purely bottom up model that only uses the likelihood terms
to infer the distribution over . This has wKfix z ]{ _hg x Cz H{ rji where i is a
normalization factor. Figure 3B shows the representational performance of this
model, through a scatter-plot of wkfZx z { against the exact posterior wyx z [ { . If
bottom-up inference was correct, then all the points would lie on the line of equality ? the bow-shape shows that purely bottom-up inference is relatively poor. Figure 4C shows this in a different way, indicating the difference between the average
summed log probabilities of the actual states under the bottom up model and those
under the true posterior. The larger and more negative the difference, the worse
the approximate inference. Averaging over l hZhih runs, the difference is m_4n h log
units (compared with a total log likelihood under the exact model of mpoql h ).
3 ACh Inference Model
Figure 2C shows the ACh-based approximate inference model. The information
about the context comes in the form of two quantities: ]\^ , the approximated contextual state having seen [ ]\^ , and ]\ ^ , which is the measure of uncertainty in that
contextual state. The idea is that ]\ ^ is reported by ACh, and is used to control
(indicated by the filled-in ellipse) the extent to which top-down information based
on ]\^ is used to influence inference about . If we were given the full exact
posterior distribution wyx ]\^ c ]\^ z [ ]\ ^ { , then one natural definition for this ACh
signal would be the uncertainty in the most likely contextual state
]\^,_
l
lm
wyx ] \ ^`_ Xz [ ] \ ^ {
(1)
Figure 4A shows the resulting ACh signal for the case of figure 3. As expected,
ACh is generally high at times when the true state | is changing, and decreases
during the periods that | is constant. During times of change, top-down information is confusing or potentially incorrect, and so bottom-up information should
dominate. This is just the putative inferential effect of ACh.
However, the ACh signal of figure 4A was calculated assuming knowledge of the
true posterior, which is unreasonable. The model of figure 2C includes the key
approximation that the only other information from [ ]\ ^ about the state of is in
the single choice of context variable ]\ ^ . The full approximate inference algorithm
becomes
wyx]\ ^
w xCz ]\ ^
wyx uc z ]\
^
wyx c
]\^u{_
r
tlm ]\ ^l m oqp [ ]\^ approximation (2)
prior over
(3)
l wyx]\^!_ ]\^ { k lul m
]\^u{_ wyxCz ]\
^ ]\^u{ s l m tum
propagation to
(4)
z [ { wyx c z ]\^ ]\ ^ { wyx z { conditioning
(5)
wyx z [ {_ l wyx c _ z [ {
marginalization
(6)
wyx z [ {_ t w x _ c z [ {
marginalization
(7)
_ argmaxl wyx _ Xz [ {
contextual inference (8)
_ lm !"Al wyx _ Xz [ {
ACh level
(9)
u
c
]
\
4
^
c
]
\
^
where
are used as approximate sufficient statistics for [ , the number of states is
t (here
t _ g ), $#&% is the Kronecker delta, and the constant of
]\^u{_
]\ ^
proportionality in equation 5 normalizes the full conditional distribution. The last
two lines show the information that is propagated to the next time step; equation 6
shows the representational answer from the model, the distribution over given
[ . These computations are all local and straightforward, except for the representation and normalization of the joint distribution over and , a point to which
we return later. Crucially, ACh exerts its influence through equation 2. If ]\ ^ is
high, then the input stimulus controlled, likelihood term dominates in the conditioning process (equation 5); if ]\^ is low, then temporal context ( ]\^ ) and the
likelihood terms balance.
One potentially dangerous aspect of this inference procedure is that it might get
unreasonably committed to a single state ]\^ _ _Uddd because it does not represent explicitly the probability accorded to the other possible values of ]\^ given
[ ]\^ . A natural way to avoid this is to bound the ACh level from below by a constant, , making approximate inference slightly more stimulus-bound than exact
inference. This approximation should add robustness. In practice, rather than use
equation 9, we use
'
$_
'(lm' lm l wyx ._
z [ ){
(10)
1
!#"%$'&( )+*
0.5
0
1
-,.,./10 2345
6789 &
0.5
0
0
?10
?30
?50
?70
0
50
100
150
200
250
300
0.1
0.2
350
400
@BA 0CE#"
D GFIH.*KJ
@BA 0CLM-"GFNH'*KJ
0.3
0.4
:23;'
238<=>0.5
:0.6
?
0.7
0.8
0.9
1
Figure 4: ACh model. A) ACh level from the exact posterior for one run. B) ACh level
- ? in the approximate model in the same run. Note the coarse similarity between A and
B. C) Solid: the mean extra representational cost for the true state 0 ? ` over that in the exact
posterior using the ACh model as a function of the minimum allowed ACh level O . Dashed:
the same quantity for the pure bottom-up model (which is equivalent to the approximate
model for O`@ ' ). Errorbars (which are almost invisible) show standard errors of the means
over ' IIfI trials.
Figure 4B shows the approximate ACh level for the same case as figure 4A, using
_ h d l . Although the detailed value of this signal is clearly different from that
arising from an exact knowledge of the posterior probabilities (in figure 4A), the
gross movements are quite similar. Note the effect of in preventing the ACh level
from dropping to h . Figure 3C shows that the ACh-based approximate posterior
values w x z [ { are much closer to the true values than for the purely bottom-up
model, particularly for values of wyx z [ ]{ near h and l , where most data lie. Figure 3F shows that inference about is noisy, but the pattern of true values | is
certainly visible. Figure 4C shows the effect of changing on the quality of inference about the true states | . This shows differences between approximate and
exact log probabilities of the true states | , averaged over l hihZh cases. If _ l , then
inference is completely stimulus-bound, just like the purely bottom-up model; values of less than h d o appear to do well for this and other settings of the parameters
of the problem. An upper bound on the performance of approximate inference can
be calculated in three steps by: i) using the exact posterior to work out and ,
ii) using these values to approximate wyx { as in equation 2, and iii) using this
approximate distribution in equation 4 and the remaining equations. The average
resulting cost (ie the average resulting difference from the log probability under
exact inference) is m j d v log units. Thus, the ACh-based approximation performs
well, and much better than purely bottom-up inference.
'
'
'
'
'
4 Discussion
We have suggested that one of the roles of ACh in cortical processing is to report
contextual uncertainty in order to control the balance between stimulus-bound,
bottom-up, processing, and contextually-bound, top-down processing. We used
the example of a hierarchical HMM in which representational inference for a middle layer should correctly reflect such a balance, and showed that a simple model
of the drive and effects of ACh leads to competent inference.
This model is clearly overly simple. In particular, it uses a localist representation
for the state , and so exact inference would be feasible. In a more realistic case,
distributed representations would be used at multiple levels in the hierarchy, and
so only one single context could be entertained at once. Then, it would also not be
possible to represent the degree of uncertainty using the level of activities of the
units representing the context, at least given a population-coded representation. It
would also be necessary to modify the steps in equations 4 and 5, since it would
be hard to represent the joint uncertainty over representations at multiple levels
in the hierarchy. Nevertheless, our model shows the feasibility of using an ACh
signal in helping propagate and use approximate information over time.
Since it is straightforward to administer cholinergic agonists and antagonists, there
are many ways to test aspects of this proposal. We plan to start by using the
paradigm of Ress et al,24 which uses fMRI techniques to study bottom-up and
top-down influences on the detection of simple visual targets. Preliminary simulation studies indicate that a hidden Markov model under controllable cholinergic
modulation can capture several aspects of existent data on animal signal detection
tasks.18
Acknowledgements
We are very grateful to Michael Hasselmo, David Heeger, Sham Kakade and Szabolcs K?ali for helpful discussions. Funding was from the Gatsby Charitable Foundation and the NSF. Reference [28] is an extended version of this paper.
References
[1] Carpenter, GA & Grossberg, S, editors (1991) Pattern Recognition by SelfOrganizing Neural Networks. Cambridge, MA: MIT Press.
[2] Dayan, P (1999). Recurrent sampling models for the Helmholtz machine. Neural Computation, 11:653-677.
[3] Dayan, P, Hinton, GE, Neal, RM & Zemel, RS (1995) The Helmholtz machine.
Neural Computation 7:889-904.
[4] Dayan, P, Kakade, S & Montague, PR (2000). Learning and selective attention.
Nature Neuroscience, 3:1218-1223.
[5] Doya, K (1999) Metalearning, neuromodulation and emotion. The 13th Toyota
Conference on Affective Minds, 46-47.
[6] Everitt, BJ & Robbins, TW (1997) Central cholinergic systems and cognition.
Annual Review of Psychology 48:649-684.
[7] Fellous, J-M, Linster, C (1998) Computational models of neuromodulation.
Neural Computation 10:771-805.
[8] Grenander, U (1976-1981) Lectures in Pattern Theory I, II and III: Pattern Analysis, Pattern Synthesis and Regular Structures. Berlin:Springer-Verlag.
[9] Hasselmo, ME (1995) Neuromodulation and cortical function: Modeling the
physiological basis of behavior. Behavioural Brain Research 67:1-27.
[10] Hasselmo, M (1999) Neuromodulation: acetylcholine and memory consolidation. Trends in Cognitive Sciences 3:351-359.
[11] Hasselmo, ME & Bower, JM (1993) Acetylcholine and memory. Trends in Neurosciences 16:218-222.
[12] Hasselmo, ME, Wyble, BP & Wallenstein, GV (1996) Encoding and retrieval
of episodic memories: Role of cholinergic and GABAergic modulation in the
hippocampus. Hippocampus 6:693-708.
[13] Hinton, GE, & Ghahramani, Z (1997) Generative models for discovering
sparse distributed representations. Philosophical Transactions of the Royal Society of London. B352:1177-1190.
[14] Holland, PC (1997) Brain mechanisms for changes in processing of conditioned stimuli in Pavlovian conditioning: Implications for behavior theory.
Animal Learning & Behavior 25:373-399.
[15] Holland, PC & Gallagher, M (1999) Amygdala circuitry in attentional and representational processes. Trends In Cognitive Sciences 3:65-73.
[16] Kakade, S & Dayan, P (2000). Dopamine bonuses. In TK Leen, TG Dietterich
& V Tresp, editors, NIPS 2000.
[17] MacKay, DM (1956) The epistemological problem for automata. In CE Shannon & J McCarthy, editors, Automata Studies. Princeton, NJ: Princeton University Press, 235-251.
[18] McGaughy, J, Kaiser, T, & Sarter, M. (1996). Behavioral vigilance following
infusions of 192 IgG=saporin into the basal forebrain: selectivity of the behavioral impairment and relation to cortical AChE-positive fiber density. Behavioral Neuroscience 110: 247-265.
[19] Mumford, D (1994) Neuronal architectures for pattern-theoretic problems.
In C Koch & J Davis, editors, Large-Scale Theories of the Cortex. Cambridge,
MA:MIT Press, 125-152.
[20] Pearce, JM & Hall, G (1980) A model for Pavlovian learning: Variation in
the effectiveness of conditioned but not unconditioned stimuli. Psychological
Review 87:532-552.
[21] Pfluger, HJ (1999) Neuromodulation during motor development and behavior. Current Opinion in Neurobiology 9:683-689.
[22] Rabiner, LR (1989) A tutorial on hidden Markov models and selected applications in speech recognition. Proceedings of the IEEE 77:257-286.
[23] Rao, RPN & Ballard, DH (1997) Dynamic model of visual recognition predicts
neural response properties in the visual cortex. Neural Computation 9:721-763.
[24] Ress, D, Backus, BT & Heeger, DJ (2000) Activity in primary visual cortex
predicts performance in a visual detection task. Nature Neuroscience 3:940-945.
[25] Sarter, M, Bruno, JP (1997) Cognitive functions of cortical acetylcholine: Toward a unifying hypothesis. Brain Research Reviews 23:28-46.
[26] Schultz, W (1998) Predictive reward signal of dopamine neurons. Journal of
Neurophysiology 80:1?27.
[27] Schultz, W, Dayan, P & Montague, PR (1997). A neural substrate of prediction
and reward. Science, 275, 1593-1599.
[28] Yu, A & Dayan, P (2002). Acetylcholine in cortical inference. Submitted to
Neural Networks.
| 2103 |@word neurophysiology:1 trial:1 version:1 middle:1 hippocampus:5 proportionality:1 gfih:1 crucially:2 propagate:1 simulation:1 r:1 solid:1 offending:1 series:1 bc:1 past:2 existing:1 current:1 contextual:9 scatter:1 written:1 readily:1 visible:1 realistic:1 wx:1 plasticity:6 shape:1 gv:1 motor:1 plot:2 rpn:1 generative:11 instantiate:2 discovering:1 nq:2 selected:1 farther:1 lr:1 unfamiliarity:1 characterization:1 coarse:1 persistent:1 incorrect:1 affective:1 behavioral:3 manner:1 expected:1 behavior:4 xz:3 brain:3 resolve:1 actual:6 jm:2 becomes:1 bonus:1 interpreted:1 substantially:2 nj:1 temporal:3 certainty:1 rm:1 uk:2 control:8 unit:8 appear:1 positive:1 understood:1 local:1 modify:1 consequence:1 encoding:1 modulation:3 might:1 contextually:1 bi:1 averaged:1 grossberg:1 arguing:2 practice:1 procedure:1 episodic:1 area:2 drug:1 wrinkle:1 oqp:1 projection:1 inferential:2 pre:1 regular:1 get:1 convenience:1 ga:1 pertain:1 storage:1 context:11 influence:5 seminal:1 twodimensional:1 equivalent:2 map:1 straightforward:3 attention:2 automaton:2 pure:1 attraction:1 dominate:1 his:1 population:1 notion:1 variation:1 analogous:1 controlling:1 play:1 hierarchy:2 target:1 exact:20 substrate:1 us:4 hypothesis:1 trend:3 helmholtz:2 recognition:10 particularly:2 approximated:1 neurobiologically:1 szabolcs:1 predicts:2 bottom:21 role:5 capture:3 readout:1 decrease:1 highest:1 movement:1 backus:1 rq:1 substantial:1 gross:1 ui:1 reward:2 dynamic:4 existent:2 weakly:1 argmaxl:1 grateful:1 ali:1 predictive:1 purely:6 completely:1 basis:1 joint:3 montague:2 represented:2 various:1 fiber:1 lul:1 describe:1 london:2 zemel:1 sgt:1 whose:1 quite:2 larger:1 valued:1 serotonin:1 favor:1 statistic:1 itself:1 noisy:1 unconditioned:1 sequence:5 grenander:1 ucl:2 propose:2 adaptation:1 bow:1 representational:7 differentially:1 ach:44 produce:1 incremental:1 tk:1 help:1 illustrate:2 recurrent:5 ac:2 paying:1 come:2 indicate:1 direction:2 thick:1 correct:1 filter:1 stochastic:2 centered:1 opinion:1 investigation:1 preliminary:1 helping:1 sufficiently:1 koch:1 hall:1 presumably:1 mapping:2 viterbi:1 bj:1 cognition:1 circuitry:1 lose:1 currently:2 individually:1 hasselmo:7 robbins:1 mit:2 clearly:2 inhibitor:1 gaussian:2 rather:3 avoid:1 hj:1 acetylcholine:6 conjunction:1 release:1 focus:1 likelihood:9 suppression:1 helpful:1 inference:34 dayan:8 synonymous:1 weakening:1 bt:1 basalis:1 hidden:5 relation:1 selective:1 fhg:1 interested:1 development:1 animal:5 plan:1 summed:1 mackay:1 uc:1 emotion:1 once:1 having:1 sampling:1 yu:2 unsupervised:1 fmri:1 report:4 stimulus:11 h_:2 simultaneously:1 individual:2 attractor:1 detection:3 izj:1 certainly:1 cholinergic:9 mixture:1 pc:2 sarter:2 hg:1 wc1n:1 implication:1 accurate:1 closer:1 necessary:1 experience:2 filled:1 theoretical:1 uncertain:1 psychological:1 increased:1 instance:1 modeling:1 rao:1 ar:1 queen:1 tg:1 cost:2 ca3:3 deviation:2 localist:1 uniform:1 stored:1 reported:3 answer:1 combined:2 density:2 ie:2 stay:1 probabilistic:1 enhance:1 synthesis:2 michael:1 ambiguity:1 neuromodulators:3 nm:1 reflect:1 central:1 vigilance:1 worse:2 stochastically:1 corner:2 cognitive:3 return:1 rji:1 includes:1 explicitly:1 later:2 view:1 start:1 square:3 formed:1 characteristic:1 rabiner:1 igg:1 plastic:1 agonist:1 drive:1 submitted:1 synapsis:1 influenced:2 metalearning:1 synaptic:2 ed:8 definition:1 against:1 colleague:1 mcgaughy:1 conveys:1 dm:1 associated:5 attributed:1 propagated:2 knowledge:2 actually:1 focusing:1 tum:1 response:1 leen:1 just:3 propagation:1 quality:2 indicated:1 effect:8 dietterich:1 true:10 equality:1 read:3 neal:1 eg:2 during:6 davis:1 rat:1 hippocampal:1 antagonist:1 theoretic:1 invisible:1 performs:1 ef:1 funding:1 conditioning:6 jp:1 extend:1 association:1 interpretation:1 discussed:1 cambridge:2 everitt:1 bruno:1 dj:1 pq:2 cortex:4 similarity:1 add:1 posterior:15 mccarthy:1 recent:1 showed:1 forcing:1 scenario:1 selectivity:1 verlag:1 epistemological:1 rotationally:1 seen:1 minimum:1 determine:2 paradigm:1 period:1 paradigmatic:1 signal:12 ud:1 dashed:1 full:3 ii:2 multiple:2 infer:1 sham:1 england:1 characterized:1 retrieval:1 permitting:1 equally:1 coded:1 controlled:1 impact:1 prediction:2 involving:1 calculates:1 feasibility:1 essentially:1 expectation:2 enhancing:1 dopamine:4 exerts:1 histogram:3 represent:4 cz:6 normalization:2 cell:1 proposal:4 wealth:1 modality:1 appropriately:1 extra:2 wallenstein:1 unlike:1 effectiveness:1 near:1 backwards:1 iii:2 variety:1 switch:1 marginalization:2 timesteps:1 psychology:1 architecture:1 ifi:1 reduce:1 idea:4 intensive:1 s7:2 septum:1 peter:1 speech:1 action:2 impairment:1 generally:1 detailed:1 selforganizing:1 concentrated:1 processed:1 category:1 documented:1 generate:1 nsf:1 tutorial:1 neuroscience:5 estimated:1 delta:2 arising:1 correctly:1 overly:1 broadly:1 discrete:1 dropping:1 basal:1 key:2 nevertheless:1 changing:3 ce:2 run:4 inverse:1 uncertainty:14 reporting:2 almost:1 doya:1 putative:2 confusing:1 bound:7 layer:8 hi:2 annual:1 activity:4 strength:1 occur:1 dangerous:1 kronecker:1 bp:1 invertebrate:1 aspect:4 pavlovian:2 relatively:1 according:2 poor:2 feraina:1 across:1 slightly:1 kakade:3 tw:1 making:2 presently:1 invariant:1 pr:2 behavioural:1 equation:9 previously:1 forebrain:1 fail:1 mechanism:2 neuromodulation:5 mind:1 ge:2 gaussians:4 permit:1 unreasonable:1 accorded:1 hierarchical:4 generic:1 robustness:1 angela:1 top:15 include:1 unreasonably:1 completed:1 remaining:1 res:2 unifying:1 infusion:1 ghahramani:1 build:1 ellipse:1 classical:1 comparatively:1 society:1 quantity:2 kaiser:1 mumford:1 primary:1 devote:1 attentional:3 mapped:1 link:1 berlin:1 hmm:3 me:3 collected:1 extent:1 toward:1 assuming:1 kalman:1 length:1 useless:1 ratio:1 balance:6 fe:1 potentially:2 negative:1 ba:3 suppress:2 allowing:1 upper:3 vertical:1 neuron:2 markov:4 pearce:1 parietal:1 extended:1 hinton:2 committed:1 neurobiology:1 wyx:22 introduced:1 david:1 specified:1 extensive:2 connection:4 philosophical:1 errorbars:1 established:1 boost:1 nip:1 suggested:3 bar:1 usually:2 pattern:9 below:1 royal:1 memory:8 overlap:3 treated:1 natural:2 advanced:1 representing:1 yq:1 gabaergic:1 auto:1 naive:1 tresp:1 kj:2 prior:6 understanding:1 sg:1 acknowledgement:1 review:3 lecture:1 foundation:1 nucleus:1 degree:1 basin:1 consistent:1 sufficient:1 principle:1 editor:4 charitable:1 linster:1 normalizes:1 consolidation:1 last:1 implicated:1 cortico:1 understand:1 wide:1 face:1 sparse:1 distributed:2 calculated:3 cortical:10 transition:3 amygdala:1 rich:1 sensory:1 forward:2 preventing:1 schultz:2 transaction:1 approximate:19 norepinephrine:1 why:1 nature:2 ballard:1 controllable:1 operational:1 cl:1 noise:1 allowed:1 competent:1 carpenter:1 neuronal:1 gatsby:4 slow:1 inferring:1 heeger:2 lie:2 governed:1 bower:1 toyota:1 down:15 e4:1 fellous:1 specific:4 showing:1 symbol:1 physiological:1 evidence:1 dominates:2 modulates:1 gallagher:1 entertained:1 conditioned:2 electrophysiology:1 likely:5 visual:5 holland:3 springer:1 dh:1 ma:2 conditional:1 identity:1 feasible:1 change:3 hard:1 determined:2 except:1 averaging:1 total:1 experimental:1 shannon:1 indicating:1 latter:1 frontal:1 evaluate:1 princeton:2 |
1,212 | 2,104 | Fast Parameter Estimation
Using Green's Functions
K. Y. Michael Wong
Department of Physics
Hong Kong University
of Science and Technology
Clear Water Bay, Hong Kong
[email protected]
FuIi Li
Department of Applied Physics
Xian Jiaotong University
Xian , China 710049
flli @xjtu. edu. en
Abstract
We propose a method for the fast estimation of hyperparameters
in large networks, based on the linear response relation in the cavity method, and an empirical measurement of the Green's function. Simulation results show that it is efficient and precise, when
compared with cross-validation and other techniques which require
matrix inversion.
1
Introduction
It is well known that correct choices of hyperparameters in classification and regression tasks can optimize the complexity of the data model , and hence achieve the
best generalization [1]. In recent years various methods have been proposed to estimate the optimal hyperparameters in different contexts, such as neural networks [2],
support vector machines [3, 4, 5] and Gaussian processes [5]. Most of these methods are inspired by the technique of cross-validation or its variant, leave-one-out
validation. While the leave-one-out procedure gives an almost unbiased estimate
of the generalization error, it is nevertheless very tedious. Many of the mentioned
attempts aimed at approximating this tedious procedure without really having to
sweat through it. They often rely on theoretical bounds, inverses to large matrices,
or iterative optimizations.
In this paper, we propose a new approach to hyperparameter estimation in large
systems. It is known that large networks are mean-field systems, so that when one
example is removed by the leave-one-out procedure, the background adjustment
can be analyzed by a self-consistent perturbation approach. Similar techniques
have been applied to the neural network [6], Bayesian learning [7] and the support
vector machine [5]. They usually involve a macroscopic number of unknown variables, whose solution is obtained through the inversion of a matrix of macroscopic
size, or iteration. Here we take a further step to replace it by a direct measurement
of the Green's function via a small number of learning processes. The proposed
procedure is fast since it does not require repetitive cross-validations, matrix inversions, nor iterative optimizations for each set of hyperparaemters. We will also
present simulation results which show that it is an excellent approximation.
The proposed technique is based on the cavity method, which was adapted from
disordered systems in many-body physics. The basis of the cavity method is a
self-consistent argument addressing the situation of removing an example from the
system. The change on removing an example is described by the Green's function,
which is an extremely general technique used in a wide range of quantum and
classical problems in many-body physics [8]. This provides an excellent framework
for the leave-one-out procedure. In this paper, we consider two applications of
the cavity method to hyperparameter estimation, namely the optimal weight decay
and the optimal learning time in feedforward networks. In the latter application,
the cavity method provides, as far as we are aware of, the only estimate of the
hyperparameter beyond empirical stopping criteria and brute force cross-validation.
2
Steady-State Hyperparameter Estimation
Consider the network with adjustable parameters w. An energy function E is
defined with respect to a set of p examples with inputs and outputs respectively
given by {IL and y'", JL = 1, ... ,p, where (IL is an N-dimensional input vector with
components
j = 1,? ?? ,N, and N ? 1 is macroscopic. We will first focus on
the dynamics of a single-layer feedforward network and generalize the results to
multilayer networks later. In single-layer networks, E has the form
e;,
E =
L f(X'",y'") + R(w).
(1)
'"
Here f( x'" , y'") represents the error function with respect to example JL. It is expressed in terms of the activation x'" == w? (IL. R( w) represents a regularization
term which is introduced to limit the complexity of the network and hence enhance
the generalization ability. Learning is achieved by the gradient descent dynamics
dWj(t) _ _ ~_oE_
dt
(2)
The time-dependent Green's function Gjk(t, s) is defined as the response of the
weight Wj at time t due to a unit stimulus added at time s to the gradient term with
respect to weight Wk, in the limit of a vanishing magnitude of the stimulus. Hence
if we compare the evolution of Wj(t) with another system Wj(t) with a continuous
perturbative stimulus Jhj(t), we would have
dWj(t) = _~ oE
dt
Now.
Jh()
J t ,
(3)
dsGjk(t,s)Jhk(s).
(4)
J
+
and the linear response relation
Wj(t) = Wj(t)
+L
J
k
Now we consider the evolution ofthe network w;'"(t) in which example JL is omitted
from the training set. For a system learning macroscopic number of examples, the
changes induced by the omission of an example are perturbative, and we can assume
that the system has a linear response. Compared with the original network Wj(t),
the gradient of the error of example JL now plays the role of the stimulus in (3).
Hence we have
(5)
Multiplying both sides by ~f and summing over j, we obtain
1-'( ) -
h t - x
I-'()
t
+
J
ds
[1
' " I-'G ( ) I-']OE(XI-'(S)'YI-')
N "7:~j jk t ,s ~k
oxl-'(s)'
(6)
Here hl-'(t) == V;\I-'(t) . ~ is called the cavity activation of example ft. When the
dynamics has reached the steady state, we arrive at
I-'
hI-'
= x
where, = limt--+oo
+,
OE(XI-' , yl-')
oxl-'
'
JdS[Ljk ~fGjk (t , s)~r]jN
(7)
is the susceptibility.
At time t , the generalization error is defined as the error function averaged over the
distribution of input (, and their corresponding output y, i.e. ,
(8)
where x == V; . (is the network activation. The leave-one-out generalization error is
an estimate of 109 given in terms ofthe cavity activations hI-' by fg = LI-' 10 (hI-' ,yl-')jp.
Hence if we can estimate the Green's function, the cavity activation in (7) provides
a convenient way to estimate the leave-one-out generalization error without really
having to undergo the validation process.
While self-consistent equations for the Green's function have been derived using
diagrammatic methods [9], their solutions cannot be computed except for the specific case of time-translational invariant Green's functions , such as those in Adaline
learning or linear regression. However, the linear response relation (4) provides a
convenient way to measure the Green's function in the general case. The basic idea
is to perform two learning processes in parallel, one following the original process
(2) and the other having a constant stimulus as in (3) with 6hj (t) = TJ6jk, where
8j k is the Kronecka delta. When the dynamics has reached the steady state, the
measurement Wj - Wj yields the quantity TJ Lk dsGjk(t, s).
J
A simple averaging procedure, replacing all the pairwise measurements between the
stimulation node k and observation node j, can be applied in the limit of large
N. We first consider the case in which the inputs are independent and normalized,
i.e., (~j) = 0, (~j~k) = 8j k. In this case, it has been shown that the off-diagonal
Green's functions can be neglected, and the diagonal Green's functions become selfaveraging, i.e. , Gjk(t , s) = G(t, s)8jk , independent of the node labels [9], rendering
, = limt--+oo J dsG(t, s).
In the case that the inputs are correlated and not normalized, we can apply standard
procedures of whitening transformation to make them independent and normalized
[1]. In large networks, one can use the diagrammatic analysis in [9] to show that
the (unknown) distribution of inputs does not change the self-averaging property of
the Green's functions after the whitening transformation. Thereafter, the measurement of Green's functions proceeds as described in the simpler case of independent
and normalized inputs. Since hyperparameter estimation usually involves a series
of computing fg at various hyperparameters, the one-time preprocessing does not
increase the computational load significantly.
Thus the susceptibility, can be measured by comparing the evolution of two processes: one following the original process (2), and the other having a constant
stimulus as in (3) with 8h j (t) = TJ for all j. When the dynamics has reached the
steady state, the measurement (Wj - Wj) yields the quantity TJ,.
We illustrate the extension to two-layer networks by considering the committee machine, in which the errorfunction takes the form E(2:: a !(x a), y) , where a = 1,? ??, nh
is the label of a hidden node, Xa == wa . [is the activation at the hidden node a,
and! represents the activation function. The generalization error is thus a function
of the cavity activations of the hidden nodes, namely, E9 = 2::JL E(2::a !(h~), yJL) /p,
where h~ = w~JL . (IL . When the inputs are independent and normalized, they are
related to the generic activations by
hJL- JL+'"
aE(2::c !(X~) , yJL)
a - Xa ~ "lab
a JL
'
Xb
b
(9)
where "lab = limt~ oo J dsGab(t, s) is the susceptibility tensor. The Green's function
Gab(t, s) represents the response of a weight feeding hidden node a due to a stimulus
applied at the gradient with respect to a weight feeding node b. It is obtained by
monitoring nh + 1 learning processes, one being original and each of the other nh
processes having constant stimuli at the gradients with respect to one of the hidden
nodes, viz.,
dw~~) (t) _
dt
1
aE
- - N ------=:(b)
aW aj
+ 'f)rSab ,
(10)
b = 1, ... ,nh?
When the dynamics has reached the steady state, the measurement (w~7
yields the quantity 'f)'Yab.
-
Waj)
We will also compare the results with those obtained by extending the analysis of
linear unlearning leave-one-out (LULOO) validation [6]. Consider the case that the
regularization R(w) takes the form of a weight decay term, R(w) = N 2::ab AabWa .
Wb/2. The cavity activations will be given by
hJL = JL + '"
a Xa ~
b
(
1-
,"
11
iJ 2:: j k ~'j(A + Q)~}bk~r
) aE(2:: c !(xn, yJL))
a JL
'
2::cjdk ~'j !'(xn(A + Q)~, dd'(x~)~r
Xb
1
(11)
where E~ represents the second derivative of E with respect to the student output
for example /1, and the matrix Aaj,bk = AabrSjk and Q is given by
Qaj,bk
= ~ 2: ~'j f'(x~)f'(x~)~r?
(12)
JL
The LULOO result of (11) differs from the cavity result of (9) in that the susceptibility "lab now depends on the example label /1, and needs to be computed by
inverting the matrix A + Q. Note also that second derivatives of the error term
have been neglected.
To verify the proposed method by simulations, we generate examples from a noisy
teacher network which is a committee machine
yJL = ~ erf
nh
(1yf2Ba ? f ) +
(Jzw
(13)
Here Ba is the teacher vector at the hidden node a. (J is the noise level. ~'j and
zJL are Gaussian variables with zero means and unit variances. Learning is done by
the gradient descent of the energy function
(14)
and the weight decay parameter ,X is the hyperparameter to be optimized. The
generalization error fg is given by
where the averaging is performed over the distribution of input { and noise z. It can
be computed analytically in terms of the inner products Qab = wa . Wb, Tab = Ba . Bb
and Rab = Ba . Wb [10]. However, this target result is only known by the teacher ,
since Tab and Rab are not accessible by the student.
Figure 1 shows the simulation results of 4 randomly generated samples. Four results
are compared: the target generalization error observed by the teacher, and those
estimated by the cavity method, cross-validation and extended LULOO. It can
be seen that the cavity method yields estimates of the optimal weight decay with
comparable precision as the other methods.
For a more systematic comparison, we search for the optimal weight decay in 10 samples using golden section search [11] for the same parameters as in Fig. 1. Compared
with the target results, the standard deviations of the estimated optimal weight decays are 0.3, 0.25 and 0.24 for the cavity method, sevenfold cross-validation and
extended LULOO respectively. In another simulation of 80 samples of the singlelayer perceptron, the estimated optimal weight decays have standard deviations of
1.2, 1.5 and 1.6 for the cavity method, tenfold cross-validation and extended LULOO respectively (the parameters in the simulations are N = 500, p = 400 and a
ranging from 0.98 to 2.56).
To put these results in perspective, we mention that the computational resources
needed by the cavity method is much less than the other estimations. For example,
in the single-layer perceptrons, the CPU time needed to estimate the optimal weight
decay using the golden section search by the teacher, the cavity method, tenfold
cross-validation and extended LULOO are in the ratio of 1 : 1.5 : 3.0 : 4.6.
Before concluding this section, we mention that it is possible to derive an expression
of the gradient dEg I d,X of the estimated generalization error with respect to the
weight decay. This provides us an even more powerful tool for hyperparameter
estimation. In the case of the search for one hyperparameter, the gradient enables
us to use the binary search for the zero of the gradient, which converges faster
than the golden section search. In the single-layer experiment we mentioned, its
precision is comparable to fivefold cross-validations, and its CPU time is only 4%
more than the teacher's search. Details will be presented elsewhere. In the case of
more than one hyperparameters, the gradient information will save us the need for
an exhaustive search over a multidimensional hyperparameter space.
3
Dynamical Hyperparameter Estimation
The second example concerns the estimation of a dynamical hyperparameter,
namely the optimal early stopping time, in cases where overtraining may plague
the generalization ability at the steady state. In perceptrons, when the examples
are noisy and the weight decay is weak, the generalization error decreases in the
early stage of learning, reaches a minimum and then increases towards its asymptotic value [12, 9]. Since the early stopping point sets in before the system reaches
the steady state, most analyses based on the equilibrium state are not applicable.
Cross-validation stopping has been proposed as an empirical method to control
overtraining [13]. Here we propose the cavity method as a convenient alternative.
0.52
G----8 target
eQ)
(b)
G----EJ cavity
0 - 0 LULOO
<=
0
~
.!::!
0.46
m
Q)
<=
Q)
0>
0.40
(d)
(c)
eQ)
<=
0
~
.!::!
m
Q)
<=
Q)
0>
0.40
o
0
weight decay A
2
weight decay A
Figure 1: (a-d) The dependence ofthe generalization error of the multilayer perceptron on the weight decay for N = 200, p = 700, nh = 3, (J = 0.8 in 4 samples. The
solid symbols locate the optimal weight decays estimated by the teacher (circle), the
cavity method (square), extended LULOO (diamond) and sevenfold cross-validation
(triangle) .
In single-layer perceptrons, the cavity activations of the examples evolve according
to (6), enabling us to estimate the dynamical evolution of the estimated generalization error when learning proceeds. The remaining issue is the measurement of
the time-dependent Green's function. We propose to introduce an initial homogeneous stimulus, that is, Jhj (t) = 1]J(t) for all j. Again, assuming normalized and
independent inputs with (~j) = 0 and (~j~k) = Jjk , we can see from (4) that the
measurement (Wj(t) - Wj(t)) yields the quantity 1]G(t, 0).
We will first consider systems that are time-translational invariant, i.e., G(t, s) =
G(t - s, 0). Such are the cases for Adaline learning and linear regression [9], where
the cavity activation can be written as
h'"(t) = x'"(t) +
J
dsG(t - s, 0) OE(X'"(S), y'").
ox,"(s)
(16)
This allows us to estimate the generalization error Eg(t) via Eg(t)
2:.," E(h'"(t), y'")/p, whose minimum in time determines the early stopping point.
To verify the proposed method in linear regression, we randomly generate examples from a noisy teacher with y'" = iJ . f'" + (Jzw Here iJ is the teacher vector with B2 = 1.
and z'" are independently generated with zero means and
unit variances. Learning is done by the gradient descent of the energy function
E(t) = 2:.," (y'" - w(t) . f'")2/2 . The generalization error Eg(t) is the error av-
e;
eraged over the distribution of input [ and their corresponding output y, i.e.,
Eg(t) = ((iJ . [ + (JZ - w? [)2/2). As far as the teacher is concerned, Eg(t) can
be computed as Eg(t) = (1 - 2R(t) + Q(t) + (J2)/2. where R(t) = w(t) . iJ and
Q(t) = W(t)2.
Figure 2 shows the simulation results of 6 randomly generated samples. Three results are compared: the teacher's estimate, the cavity method and cross-validation.
Since LULOO is based on the equilibrium state, it cannot be used in the present
context. Again, we see that the cavity method yields estimates of the early stopping time with comparable precision as cross-validation. The ratio of the CPU time
between the cavity method and fivefold cross-validation is 1 : 1.4.
For nonlinear regression and multilayer networks, the Green 's functions are not
time-translational invariant. To estimate the Green 's functions in this case, we have
devised another scheme of stimuli. Preliminary results for the determination of the
early stopping point are satisfactory and final results will be presented elsewhere.
1 .1
eQ.i
c::
a
~
.!::!
0.9
~
<l>
c::
<l>
0>
0.7
eQ.i
c::
a
~
c;;
Q.i
.!::!
0.9
c::
<l>
0>
0.7
0
2
time t
0
2
time t
0
2
4
time t
Figure 2: (a-f) The evolution of the generalization error of linear regression for
N = 500, p = 600 and (J = 1. The solid symbols locate the early stopping points
estimated by the teacher (circle), the cavity method (square) and fivefold crossvalidation (diamond).
4
Conclusion
We have proposed a method for the fast estimation of hyperparameters in large
networks, based on the linear response relation in the cavity method, combined
with an empirical method of measuring the Green's function. Its efficiency depends
on the independent and identical distribution of the inputs, greatly reducing the
number of networks to be monitored. It does not require the validation process
or the inversion of matrices of macroscopic size, and hence its speed compares
favorably with cross-validation and other perturbative approaches such as extended
LULOO. For multilayer networks, we will explore further speedup of the Green 's
function measurement by multiplexing the stimuli to the different hidden units into
a single network, to be compared with a reference network. We will also extend the
technique to other benchmark data to study its applicability.
Our initial success indicates that it is possible to generalize the method to more
complicated systems in the future. The concept of Green's functions is very general,
and its measurement by comparing the states of a stimulated system with a reference
one can be adopted to general cases with suitable adaptation. Recently, much
attention is paid to the issue of model selection in support vector machines [3, 4, 5].
It would be interesting to consider how the proposed method can contribute to these
cases.
Acknowledgements
We thank C. Campbell for interesting discussions and H. Nishimori for encouragement. This work was supported by the grant HKUST6157/99P from the Research
Grant Council of Hong Kong.
References
[1] C. M. Bishop, Neural Networks for Pattern Recognition, Clarendon Press, Oxford
(1995).
[2] G. B. Orr and K-R. Muller, eds., Neural Networks: Tricks of th e Trad e, Springer,
Berlin (1998).
[3] O. Chapelle and V. N. Vapnik, Advances in Neural Information Processing Systems
12, S. A. Solla, T. KLeen and K-R. Muller, eds., MIT Press, Cambridge, 230 (2000).
[4] S. S. Keerthi, Technical Report CD-OI-02,
http://guppy.mpe.nus. edu .sg/ mpessk/nparm.html (2001).
[5] M. Opper and O. Winther , Advances in Large Margin Classifiers, A. J. Smola, P.
Bartlett, B. Sch6lkopf and D. Schuurmans, eds., MIT Press, Cambridge, 43 (1999) .
[6] J. Larsen and L. K Hansen , Advances in Computational Math ematics 5 , 269 (1996).
[7] M. Opper and O. Winther , Phys. R ev. Lett. 76 , 1964 (1996).
[8] A. L. Fetter and J. D. Walecka, Quantum Theory of Many-Particle Systems, McGrawHill, New York (1971).
[9] K Y. M. Wong, S. Li and Y. W . Tong, Phys. Rev. E 62 , 4036 (2000).
[10] D. Saad and S. A. Solla, Phys. R ev. Lett. 74, 4337 (1995).
[11] W. H. Press, B. P. Flannery, S. A. Teukolsky and W . T. Vett erling, Num erical
R ecipes in C: Th e Art of Sci entific Computing, Cambridge University Press, Cambridge (1990).
[12] A. Krogh and J. A. Hertz , J. Phys. A 25 , 1135 (1992).
[13] S. Amari, N. Murata, K-R. Muller , M. Finke and H. H. Yang, IEEE Trans. on N eural
N etworks 8, 985 (1997) .
| 2104 |@word kong:3 inversion:4 tedious:2 simulation:7 paid:1 mention:2 solid:2 initial:2 series:1 comparing:2 activation:12 perturbative:3 ust:1 written:1 kleen:1 enables:1 vanishing:1 num:1 provides:5 math:1 node:10 contribute:1 simpler:1 direct:1 become:1 unlearning:1 introduce:1 pairwise:1 nor:1 gjk:2 inspired:1 cpu:3 tenfold:2 considering:1 transformation:2 multidimensional:1 golden:3 classifier:1 brute:1 unit:4 control:1 grant:2 before:2 limit:3 oxford:1 china:1 range:1 averaged:1 differs:1 procedure:7 empirical:4 significantly:1 convenient:3 cannot:2 selection:1 put:1 context:2 wong:2 optimize:1 attention:1 independently:1 dw:1 target:4 play:1 xjtu:1 homogeneous:1 trick:1 recognition:1 jk:2 xian:2 role:1 ft:1 observed:1 wj:12 oe:4 solla:2 decrease:1 removed:1 mentioned:2 complexity:2 neglected:2 dynamic:6 hjl:2 efficiency:1 basis:1 triangle:1 various:2 fast:4 exhaustive:1 whose:2 aaj:1 amari:1 ability:2 erf:1 noisy:3 final:1 propose:4 product:1 adaptation:1 j2:1 adaline:2 achieve:1 yjl:4 crossvalidation:1 extending:1 leave:7 gab:1 converges:1 oo:3 illustrate:1 derive:1 jjk:1 measured:1 ij:5 eq:4 krogh:1 involves:1 correct:1 disordered:1 require:3 feeding:2 generalization:17 really:2 preliminary:1 extension:1 equilibrium:2 early:7 omitted:1 susceptibility:4 estimation:11 applicable:1 label:3 hansen:1 council:1 tool:1 mit:2 gaussian:2 hj:1 ej:1 derived:1 focus:1 viz:1 indicates:1 hk:1 greatly:1 dependent:2 stopping:8 hidden:7 relation:4 translational:3 classification:1 issue:2 html:1 art:1 field:1 aware:1 having:5 identical:1 represents:5 future:1 report:1 stimulus:11 randomly:3 keerthi:1 attempt:1 ab:1 analyzed:1 ematics:1 tj:3 xb:2 circle:2 erical:1 theoretical:1 wb:3 measuring:1 finke:1 applicability:1 addressing:1 deviation:2 aw:1 teacher:12 combined:1 winther:2 accessible:1 systematic:1 physic:4 yl:2 off:1 michael:1 enhance:1 again:2 e9:1 derivative:2 li:3 orr:1 student:2 wk:1 b2:1 depends:2 later:1 performed:1 lab:3 tab:2 mpe:1 reached:4 mcgrawhill:1 parallel:1 complicated:1 il:4 square:2 oi:1 variance:2 murata:1 yield:6 ofthe:3 generalize:2 weak:1 bayesian:1 multiplying:1 monitoring:1 mpessk:1 overtraining:2 reach:2 phys:4 ed:3 energy:3 selfaveraging:1 larsen:1 monitored:1 campbell:1 clarendon:1 dt:3 response:7 done:2 ox:1 xa:3 stage:1 smola:1 d:1 replacing:1 nonlinear:1 rab:2 aj:1 jhj:2 normalized:6 unbiased:1 verify:2 concept:1 evolution:5 hence:6 regularization:2 analytically:1 satisfactory:1 eg:6 self:4 steady:7 hong:3 criterion:1 ranging:1 recently:1 stimulation:1 jp:1 nh:6 jl:11 extend:1 measurement:11 cambridge:4 encouragement:1 particle:1 chapelle:1 whitening:2 recent:1 perspective:1 binary:1 success:1 yi:1 muller:3 jiaotong:1 seen:1 minimum:2 technical:1 faster:1 determination:1 cross:15 devised:1 variant:1 regression:6 basic:1 multilayer:4 ae:3 iteration:1 repetitive:1 limt:3 achieved:1 background:1 oxl:2 qab:1 macroscopic:5 saad:1 induced:1 undergo:1 yang:1 feedforward:2 concerned:1 rendering:1 fetter:1 sweat:1 idea:1 inner:1 expression:1 bartlett:1 york:1 clear:1 aimed:1 involve:1 jhk:1 yab:1 generate:2 http:1 delta:1 estimated:7 hyperparameter:11 thereafter:1 four:1 nevertheless:1 year:1 inverse:1 powerful:1 arrive:1 almost:1 comparable:3 bound:1 layer:6 hi:3 adapted:1 multiplexing:1 speed:1 argument:1 extremely:1 concluding:1 speedup:1 department:2 according:1 eraged:1 hertz:1 rev:1 fivefold:3 hl:1 invariant:3 equation:1 resource:1 etworks:1 committee:2 needed:2 adopted:1 apply:1 generic:1 save:1 alternative:1 jn:1 original:4 remaining:1 approximating:1 classical:1 tensor:1 added:1 quantity:4 dependence:1 diagonal:2 gradient:11 thank:1 berlin:1 sci:1 dwj:2 water:1 assuming:1 ratio:2 dsg:2 favorably:1 ba:3 unknown:2 adjustable:1 perform:1 sevenfold:2 diamond:2 observation:1 av:1 sch6lkopf:1 benchmark:1 enabling:1 descent:3 situation:1 extended:6 precise:1 locate:2 perturbation:1 omission:1 zjl:1 introduced:1 bk:3 namely:3 inverting:1 optimized:1 waj:1 plague:1 nu:1 guppy:1 trans:1 beyond:1 proceeds:2 usually:2 dynamical:3 pattern:1 ev:2 green:20 suitable:1 rely:1 force:1 scheme:1 technology:1 lk:1 sg:1 acknowledgement:1 singlelayer:1 evolve:1 nishimori:1 asymptotic:1 interesting:2 validation:19 consistent:3 dd:1 dsgjk:2 cd:1 elsewhere:2 supported:1 side:1 jh:1 perceptron:2 wide:1 fg:3 opper:2 xn:2 lett:2 quantum:2 preprocessing:1 far:2 bb:1 cavity:27 deg:1 summing:1 xi:2 continuous:1 iterative:2 search:8 bay:1 stimulated:1 jz:1 schuurmans:1 excellent:2 noise:2 hyperparameters:6 body:2 eural:1 fig:1 en:1 tong:1 precision:3 removing:2 diagrammatic:2 load:1 specific:1 bishop:1 phkywong:1 symbol:2 decay:14 concern:1 vapnik:1 magnitude:1 margin:1 flannery:1 explore:1 expressed:1 adjustment:1 springer:1 determines:1 teukolsky:1 ljk:1 towards:1 replace:1 change:3 except:1 reducing:1 averaging:3 called:1 perceptrons:3 support:3 latter:1 hkust6157:1 correlated:1 |
1,213 | 2,105 | Algorithmic Luckiness
Ralf Herbrich
Microsoft Research Ltd.
CB3 OFB Cambridge
United Kingdom
rherb@microsoft?com
Robert C. Williamson
Australian National University
Canberra 0200
Australia
Bob. Williamson @anu.edu.au
Abstract
In contrast to standard statistical learning theory which studies
uniform bounds on the expected error we present a framework that
exploits the specific learning algorithm used. Motivated by the
luckiness framework [8] we are also able to exploit the serendipity
of the training sample. The main difference to previous approaches
lies in the complexity measure; rather than covering all hypotheses in a given hypothesis space it is only necessary to cover the
functions which could have been learned using the fixed learning
algorithm. We show how the resulting framework relates to the
VC, luckiness and compression frameworks. Finally, we present an
application of this framework to the maximum margin algorithm
for linear classifiers which results in a bound that exploits both the
margin and the distribution of the data in feature space.
1
Introduction
Statistical learning theory is mainly concerned with the study of uniform bounds
on the expected error of hypotheses from a given hypothesis space [9, 1]. Such
bounds have the appealing feature that they provide performance guarantees for
classifiers found by any learning algorithm. However, it has been observed that
these bounds tend to be overly pessimistic. One explanation is that only in the
case of learning algorithms which minimise the training error it has been proven
that uniformity of the bounds is equivalent to studying the learning algorithm's
generalisation performance directly.
In this paper we present a theoretical framework which aims at directly studying the
generalisation error of a learning algorithm rather than taking the detour via the
uniform convergence of training errors to expected errors in a given hypothesis space.
In addition, our new model of learning allows the exploitation of the fact that we
serendipitously observe a training sample which is easy to learn by a given learning
algorithm. In that sense, our framework is a descendant of the luckiness framework
of Shawe-Taylor et al. [8]. In the present case, the luckiness is a function of a given
learning algorithm and a given training sample and characterises the diversity of
the algorithms solutions. The notion of luckiness allows us to study given learning
algorithms at many different perspectives. For example, the maximum margin
algorithm [9] can either been studied via the number of dimensions in feature space,
the margin of the classifier learned or the sparsity of the resulting classifier. Our
main results are two generalisation error bounds for learning algorithms: one for
the zero training error scenario and one agnostic bound (Section 2). We shall
demonstrate the usefulness of our new framework by studying its relation to the
VC framework, the original luckiness framework and the compression framework of
Littlestone and Warmuth [6] (Section 3). Finally, we present an application of the
new framework to the maximum margin algorithm for linear classifiers (Section 4).
The detailed proofs of our main results can be found in [5].
We denote vectors using bold face, e.g. x = (Xl, ... ,xm ) and the length of this vector
by lxi, i.e. Ixl = m. In order to unburden notation we use the shorthand notation
Z[i:jJ := (Zi,"" Zj) for i :::; j. Random variables are typeset in sans-serif font. The
symbols P x , Ex [f (X)] and IT denote a probability measure over X, the expectation of
f (.) over the random draw of its argument X and the indicator function, respectively.
The shorthand notation Z(oo) := U;;'=l zm denotes the union of all m- fold Cartesian
products of the set Z. For any mEN we define 1m C {I, ... , m }m as the set of all
permutations of the numbers 1, ... ,m,
1m := {(i l , ... ,i m) E {I, ... ,m}m I'v'j f:- k: ij f:- id .
Given a 2m- vector i E hm and a sample z E z2m we define Wi : {I, ... , 2m} -+
{I, ... , 2m} by Wi (j) := ij and IIdz) by IIi (z) := (Z7ri(l), ... , Z7ri(2m))'
2
Algorithmic Luckiness
Suppose we are given a training sample z = (x, y) E (X x y)m = zm of size
mEN independently drawn (iid) from some unknown but fixed distribution P Xy =
P z together with a learning algorithm A : Z( 00) -+ yX . For a predefined loss
l : y x y -+ [0,1] we would like to investigate the generalisation error Gl [A, z] :=
Rl [A (z)] - infhEYx Rl [h] of the algorithm where the expected error Rl [h] of his
defined by
Rl [h] := Exy [l (h (X) ,Y)] .
Since infhEYx Rl [h] (which is also known as the Bayes error) is independent of A
it suffices to bound Rl [A (z)]. Although we know that for any fixed hypothesis h
the training error
~
1
Rdh,z]:=~
L
l(h(xi),Yi)
(X i ,Yi) E z
is with high probability (over the random draw of the training sample z E Z(oo))
close to Rl [h], this might no longer be true for the random hypothesis A (z). Hence
we would like to state that with only small probability (at most 8) , the expected
error Rl [A (z)] is larger than the training error HI [A (z), z] plus some sample and
algorithm dependent complexity c (A, z, 8),
Pzm (Rl [A (Z)] > HI [A (Z), Z]
+ c (A, Z,8)) < 8.
(1)
In order to derive such a bound we utilise a modified version of the basic lemma of
Vapnik and Chervonenkis [10].
Lemma 1. For all loss functions l : y x y -+ [0,1], all probability measures P z , all
algorithms A and all measurable formulas Y : zm -+ {true, false}, if mc 2 > 2 then
Pzm (( RdA (Z)] > HdA (Z) , Z]
+ c)
2P Z 2m ((HI [A (Z[l:m]) ,Z[(m+l):2mJJ
/\ Y (Z)) <
> HI [A (Z[l:mJ) ,Z[l:mJJ + ~) /\ Y (Z[l:m])) .
,
.I
V
J(Z)
Proof (Sketch). The probability on the r.h.s. is lower bounded by the probability of the conjunction of event on the l.h.s. and Q (z)
Rl [A (Z[l:mj)] Rl [A (Z[l:mj) ,Z(m+1):2m] < ~. Note that this probability is over z E z2m. If
we now condition on the first m examples, A (Z[l:mj) is fixed and therefore by an
application of Hoeffding's inequality (see, e.g. [1]) and since m?2 > 2 the additional
event Q has probability of at least ~ over the random draw of (Zm+1, ... , Z2m). 0
=
Use of Lemma 1 - which is similar to the approach of classical VC analysis reduces the original problem (1) to the problem of studying the deviation of the
training errors on the first and second half of a double sample z E z2m of size
2m. It is of utmost importance that the hypothesis A (Z[l:mj) is always learned
from the first m examples. Now, in order to fully exploit our assumptions of the
mutual independence of the double sample Z E z2m we use a technique known
as symmetrisation by permutation: since PZ2~ is a product measure, it has the
property that PZ2?> (J (Z)) = PZ2~ (J (ITi (Z))) for any i E hm. Hence, it suffices
to bound the probability of permutations Jri such that J (ITi (z)) is true for a given
and fixed double sample z. As a consequence thereof, we only need to count the
number of different hypotheses that can be learned by A from the first m examples
when permuting the double sample.
Definition 1 (Algorithmic luckiness). Any function L that maps an algorithm
A : Z( oo ) -+ yX and a training sample z E Z( oo ) to a real value is called an algorithmic luckiness. For all mEN, for any z E z2m , the lucky set HA (L , z) ~ yX is the
set of all hypotheses that are learned from the first m examples (Z7ri(1),???, Z7ri(m))
when permuting the whole sample z whilst not decreasing the luckiness, i.e.
(2)
where
Given a fixed loss function 1 :
y
x
y -+ [0,1]
the induced loss function set
?1 (HA (L,z)) is defined by
?1 (HA (L,z)) := {(x,y) r-+ 1(h(x) ,y)
I h E HA (L,z)}
.
For any luckiness function L and any learning algorithm A , the complexity of the
double sample z is the minimal number N1 (T, ?1 (HA (L, z)) ,z) of hypotheses h E
yX needed to cover ?1 (HA (L , z)) at some predefined scale T, i.e. for any hypothesis
hE HA (L, z) there exists a h E yX such that
(4)
To see this note that whenever J (ITi (z)) is true (over the random draw of permutations) then there exists a function h which has a difference in the training errors
on the double sample of at least ~ + 2T. By an application of the union bound we
see that the number N 1 (T, ?1 (HA (L , z)) , z) is of central importance. Hence, if we
are able to bound this number over the random draw of the double sample z only
using the luckiness on the first m examples we can use this bound in place of the
worst case complexity SUPzEZ2~ N1 (T, ?1 (HA (L , z)) ,z) as usually done in the VC
framework (see [9]).
Definition 2 (w- smallness of L). Given an algorithm A : Z (00 ) -+ yX and a loss
l : y x y -+ [a, 1] the algorithmic luckiness function Lis w- small at scale T E jR+ if
for all mEN, all J E (a , 1] and all P z
PZ2~ (Nl (T, ?"1 (1iA (L, Z)), Z)
> w (L (A, Z[l:ml) ,l, m, J,T)) < J.
,
"
v
S(Z)
Note that if the range of l is {a, I} then N 1 (2~ ' ?"1 (1iA (L, z)) , z) equals the number of dichotomies on z incurred by ?"1 (1iA (L , z)).
Theorem 1 (Algorithmic luckiness bounds). Suppos e we have a learning
algorithm A : Z( oo ) -+ yX and an algorithmic luckiness L that is w-small at
scale T for a loss function l : y X Y -+ [a, 1]. For any probability measure P z ,
any dEN and any J E (a , 1], with probability at least 1 - J over the random
draw of the training sample z E zm of size m, if w (L (A, z) ,l, m, J/4, T) :::; 2d
then
!
Rz[A (z)] :::; Rz[A (z), z] +
(d + 10g2
(~) )
+ 4T.
(5)
Furthermore, under the above conditions if the algorithmic luckiness L is wsmall at scale 2~ for a binary loss function l (".) E {a, I} and Rl [A (z), z] = a
then
(6)
Proof (Compressed Sketch). We will only sketch the proof of equation (5) ; the proof
of (6) is similar and can be found in [5]. First, we apply Lemma 1 with Y (z) ==
w (L (A,z) ,l,m,J/4,T) :::; 2d. We now exploit the fact that
PZ2~
(J (Z))
:Z2~
(J (Z)
1\
S (Z) ), +PZ2~ (J (Z)
1\
...,S (Z))
v
J
< 4+
:::: P Z 2 ~
(S(Z))
PZ2~
(J (Z) I\...,S (Z)) ,
which follows from Definition 2. Following the above-mentioned argument it suffices to bound the probability of a random permutation III (z) that J (III (z)) 1\
...,S (III (z)) is true for a fixed double sample z. Noticing that Y (z) 1\ ...,S (z) =>
Nl (T,?"l (1iA (L , z)) ,z) :::; 2d we see that we only consider swappings Jri for which
Nl (T,?"l (1iA (L,IIi (z))) ,IIi (z)) :::; 2d. Thus let us consider such a cover of
size not more than 2. By (4) we know that whenever J (IIi (z)) 1\ ...,S (IIi (z))
is true for a swapping i then there exists a hypothesis h E yX in the cover
(III (z)) [(m+1) :2ml] - Rl
(III (z)) [l:ml] > ~ + 2T. Using the
such that Rl
union bound and Hoeffding's inequality for a particular choice of PI shows that
PI (J (III (z)) 1\ ...,S (III (z))) :::; ? which finalises the proof.
D
[h,
[h,
A closer look at (5) and (6) reveals that the essential difference to uniform bounds
on the expected error is within the definition of the covering number: rather than
covering all hypotheses h in a given hypothesis space 1i ~ yX for a given double
sample it suffices to cover all hypotheses that can be learned by a given learning
algorithm from the first half when permuting the double sample. Note that the
usage of permutations in the definition of (2) is not only a technical matter; it
fully exploits all the assumptions made for the training sample, namely the training
sample is drawn iid.
3
Relationship to Other Learning Frameworks
In this section we present the relationship of algorithmic luckiness to other learning
frameworks (see [9, 8, 6] for further details of these frameworks).
VC Framework If we consider a binary loss function l (".) E {a, I} and assume
that the algorithm A selects functions from a given hypothesis space H ~ yX then
L (A, z) = - VCDim (H) is a w- smallluckiness function where
w
(Lo,l,m,8, 1) :S (2em)
-Lo
2m
-Lo
.
(7)
This can easily be seen by noticing that the latter term is an upper bound on
maxz EZ 2 ", I{ (l (h (Xl) ,yI) , ... ,l (h (X2m), Y2m)) : h E H}I (see also [9]). Note that
this luckiness function neither exploits the particular training sample observed nor
the learning algorithm used.
Luckiness Framework Firstly, the luckiness framework of Shawe-Taylor et al. [8]
only considered binary loss functions l and the zero training error case. In this work,
the luckiness ? is a function of hypothesis and training samples and is called wsmall if the probability over the random draw of a 2m sample z that there exists a
hypothesis h with w(?(h, (Zl, ... ,zm )), 8) < J'--h (2;" {(X , y) t--+ l (g (x) ,y) 1? (g , z) :::::
? (h, Z)}, z), is smaller than 8. Although similar in spirit, the classical luckiness
framework does not allow exploitation of the learning algorithm used to the same
extent as our new luckiness. In fact, in this framework not only the covering number
must be estimable but also the variation of the luckiness ? itself. These differences
make it very difficult to formally relate the two frameworks.
Compression Framework In the compression framework of Littlestone and
Warmuth [6] one considers learning algorithms A which are compression schemes,
i.e. A (z) = :R (e (z)) where e (z) selects a subsample z ~ z and :R : Z(oo) -+ yX
is a permutation invariant reconstruction function. For this class of learning algorithms, the luckiness L(A,z) = -le(z)1 is w- small where w is given by (7). In
order to see this we note that (3) ensures that we only consider permutations 7ri
where e (IIi (z)) :S Ie (z)l, i.e. we use not more than -L training examples from
z E z2m. As there are exactly
distinct choices of d training examples from
2m examples the result follows by application of Sauer's lemma [9]. Disregarding
constants, Theorem 1 gives exactly the same bound as in [6].
e;;)
4
A New Margin Bound For Support Vector Machines
In this section we study the maximum margin algorithm for linear classifiers, i.e. A :
Z(oo) -+ Hcp where Hcp := {x t--+ (? (x), w) I wE }C} and ? : X -+ }C ~ ?~ is known
as the feature mapping. Let us assume that l (h (x) ,y) = lO - l (h (x) ,y) := lIyh(x)::;o,
Classical VC generalisation error bounds exploit the fact that VCDim (Hcp) = nand
(7). In the luckiness framework of Shawe-Taylor et al. [8] it has been shown that we
can use fat1i.p h'z (w)) :S h'z (W))-2 (at the price of an extra 10g2 (32m) factor) in
place of VCDim (Hcp) where "(z (w) = min(xi,Yi)Ez Yi (? (Xi) , w) / Ilwll is known as
the margin. Now, the maximum margin algorithm finds the weight vector WMM that
maximises "(z (w). It is known that WMM can be written as a linear combination of
the ? (Xi). For notational convenience, we shall assume that A: Z(oo) -+ 1R(00) maps
to the expansion coefficients 0: such that Ilwall = 1 where Wa := 2:1~1 (XicfJ(Xi).
Our new margin bound follows from the following theorem together with (6).
Theorem 2. Let fi (x) be the smallest 10 > 0 such that {cfJ (Xl) , ... , cfJ (Xm) }
can be covered by at most i balls of radius less than or equal f. Let f z (w) be
.
Yi (4)(X i), W)
D
th
l
l
d
defi ne d by f z (W ) .. - mm( Xi, Yi)Ez 1
14>(Xi) II.llwll. ror e zero-one oss 0-1 an
the maximum margin algorithm A , the luckiness function
L(A ,Z )
=_ mIn
. {.~ E
",,-T
1'1
.>
_
~
(f (X)2:7=1 IA
i
fz
(Z)jl) 2 }
( )
,
(8)
W A(z)
is w-small at scale 112m w.r.t. the function
(
1)
w L o,l,m,8, 2m
=
(2em)- 2L O
-Lo
(9)
Proof (Sketch). First we note that by a slight refinement of a theorem of Makovoz
[7] we know that for any Z E zm there exists a weight vector w = 2:: 1 iiicfJ (Xi)
such that
(10)
and
a
WA(z)
E ]Rm has no more than - L (A, z) non-zero components. Although only
is of unit length, one can show that (10) implies that
(WA(z),
wi IIwll) ~ )1- f; (WA(z?).
Using equation (10) of [4] this implies that w correctly classifies Z E zm. Consider
a fixed double sample Z E z2m and let ko := L (A, (Zl , ... , zm )). By virtue of (3)
and the aforementioned argument we only need to consider permutations tri such
that there exists a weight vector w = 2:;:1 iijcfJ (Xj) with no more than ko non-zero
iij. As there are exactly (2;;) distinct choices of dE {I, ... , ko} training examples
from the 2m examples Z there are no more than (2emlko)kO different subsamples
to be used in w. For each particular subsample z ~ Z the weight vector w is a
member of the class of linear classifiers in a ko (or less) dimensional space. Thus,
from (7) it follows that for the given subsample z there are no more (2emlko)kO
different dichotomies induced on the double sample Z E z2m. As this holds for any
D
double sample, the theorem is proven.
There are several interesting features about this margin bound. Firstly, observe
that 2:;:1 IA (Z)j I is a measure of sparsity of the solution found by the maximum
margin algorithm which, in the present case, is combined with margin. Note that
for normalised data, i.e. IlcfJ Oil = constant, the two notion of margins coincide,
i.e. f z (w) = I Z (w). Secondly, the quantity fi (x) can be considered as a measure
of the distribution of the mapped data points in feature space. Note that for all
i E N, fi (x) :S 101 (x) :S maxjE{l ,... ,m} IlcfJ (xj)ll. Supposing that the two classconditional probabilities PX 1Y=y are highly clustered, 102 (x) will be very small. An
extension of this reasoning is useful in the multi-class case; binary maximum margin
classifiers are often used to solve multi-class problems [9]. There appears to be also
a close relationship of fi (x) with the notion of kernel alignment recently introduced
in [3]. Finally, one can use standard entropy number techniques to bound fi (x) in
terms of eigenvalues of the inner product matrix or its centred variants. It is worth
mentioning that although our aim was to study the maximum margin algorithm the
above theorem actually holds for any algorithm whose solution can be represented
as a linear combination of the data points.
5
Conclusions
In this paper we have introduced a new theoretical framework to study the generalisation error of learning algorithms. In contrast to previous approaches, we
considered specific learning algorithms rather than specific hypothesis spaces. We
introduced the notion of algorithmic luckiness which allowed us to devise data dependent generalisation error bounds. Thus we were able to relate the compression
framework of Littlestone and Warmuth with the VC framework. Furthermore, we
presented a new bound for the maximum margin algorithm which not only exploits
the margin but also the distribution of the actual training data in feature space.
Perhaps the most appealing feature of our margin based bound is that it naturally combines the three factors considered important for generalisation with linear
classifiers: margin, sparsity and the distribution of the data. Further research is
concentrated on studying Bayesian algorithms and the relation of algorithmic luckiness to the recent findings for stable learning algorithms [2].
Acknowledgements This work was done while RCW was visiting Microsoft Research Cambridge. This work was also partly supported by the Australian Research
Council. RH would like to thank Olivier Bousquet for stimulating discussions.
References
[1) M. Anthony and P. Bartlett. A Theory of Learning in Artificial Neural Networks.
Cambridge University Press, 1999.
[2) O. Bousquet and A. Elisseeff. Algorithmic stability and generalization performance. In
T. K. Leen , T. G. Dietterich, and V. Tresp, editors, Advances in N eural Information
Processing Systems 13, pages 196- 202. MIT Press, 2001.
[3) N. Cristianini, A. Elisseeff, and J. Shawe-Taylor. On optimizing kernel alignment .
Technical Report NC2-TR-2001-087, NeuroCOLT, http: //www.neurocolt.com. 2001.
[4) R. Herbrich and T . Graepel. A PAC-Bayesian margin bound for linear classifiers:
Why SVMs work. In T. K. Leen, T. G. Dietterich, and V. Tresp, editors, Advances
in Neural Information Processing Systems 13, pages 224- 230 , Cambridge, MA , 2001.
MIT Press.
[5) R. Herbrich and R. C. Williamson. Algorithmic luckiness. Technical report, Microsoft
Research, 2002.
[6) N . Littlestone and M. Warmuth. Relating data compression and learnability. Technical report, University of California Santa Cruz, 1986.
[7) Y . Makovoz. Random approximants and neural networks. Journal of Approximation
Theory, 85:98- 109, 1996.
[8) J. Shawe-Taylor, P. L. Bartlett, R. C. Williamson, and M. Anthony. Structural risk
minimization over data-dependent hierarchies. IEEE Transactions on Information
Theory, 44(5):1926- 1940, 1998.
[9) V . Vapnik. Statistical Learning Theory. John Wiley and Sons, New York, 1998.
[10) V. N. Vapnik and A. Y. Chervonenkis. On the uniform convergence of relative frequencies of events to their probabilities. Theory of Probability and its Applications,
16(2):264- 281, 1971.
| 2105 |@word exploitation:2 version:1 compression:7 elisseeff:2 tr:1 united:1 chervonenkis:2 com:2 z2:1 exy:1 must:1 written:1 john:1 cruz:1 half:2 warmuth:4 herbrich:3 firstly:2 descendant:1 shorthand:2 combine:1 expected:6 os:1 nor:1 multi:2 decreasing:1 rdh:1 actual:1 classifies:1 notation:3 bounded:1 agnostic:1 whilst:1 finding:1 guarantee:1 exactly:3 classifier:10 rm:1 zl:2 unit:1 consequence:1 id:1 might:1 plus:1 au:1 studied:1 mentioning:1 range:1 union:3 cfj:2 lucky:1 convenience:1 close:2 risk:1 www:1 equivalent:1 measurable:1 map:2 maxz:1 independently:1 his:1 ralf:1 stability:1 notion:4 variation:1 hierarchy:1 suppose:1 olivier:1 hypothesis:20 defi:1 observed:2 worst:1 ensures:1 mentioned:1 complexity:4 symmetrisation:1 cristianini:1 uniformity:1 ror:1 easily:1 represented:1 distinct:2 artificial:1 dichotomy:2 whose:1 larger:1 solve:1 compressed:1 typeset:1 itself:1 subsamples:1 eigenvalue:1 reconstruction:1 product:3 zm:9 convergence:2 double:13 oo:8 derive:1 ij:2 implies:2 australian:2 radius:1 vc:7 australia:1 vcdim:3 suffices:4 clustered:1 generalization:1 rda:1 pessimistic:1 secondly:1 extension:1 sans:1 mm:1 hold:2 considered:4 algorithmic:13 mapping:1 smallest:1 council:1 minimization:1 mit:2 always:1 aim:2 modified:1 rather:4 conjunction:1 notational:1 mainly:1 contrast:2 sense:1 dependent:3 nand:1 relation:2 selects:2 aforementioned:1 mutual:1 equal:2 look:1 report:3 national:1 microsoft:4 n1:2 investigate:1 highly:1 alignment:2 nl:3 swapping:1 permuting:3 predefined:2 closer:1 necessary:1 xy:1 sauer:1 detour:1 taylor:5 littlestone:4 theoretical:2 minimal:1 approximants:1 cover:5 deviation:1 uniform:5 usefulness:1 learnability:1 combined:1 ie:1 together:2 central:1 serendipitously:1 hoeffding:2 li:1 diversity:1 de:1 centred:1 bold:1 coefficient:1 matter:1 bayes:1 nc2:1 bayesian:2 iid:2 mc:1 worth:1 bob:1 whenever:2 definition:5 frequency:1 thereof:1 naturally:1 proof:7 graepel:1 actually:1 appears:1 leen:2 done:2 furthermore:2 sketch:4 perhaps:1 oil:1 dietterich:2 usage:1 true:6 hence:3 ll:1 covering:4 demonstrate:1 reasoning:1 fi:5 recently:1 rl:14 jl:1 he:1 slight:1 relating:1 cambridge:4 shawe:5 stable:1 longer:1 recent:1 perspective:1 optimizing:1 scenario:1 inequality:2 binary:4 yi:7 devise:1 seen:1 additional:1 ii:1 relates:1 reduces:1 technical:4 hda:1 variant:1 basic:1 ko:6 expectation:1 kernel:2 addition:1 extra:1 tri:1 induced:2 tend:1 supposing:1 member:1 spirit:1 z2m:9 structural:1 iii:13 easy:1 concerned:1 independence:1 rherb:1 zi:1 xj:2 inner:1 minimise:1 motivated:1 bartlett:2 ltd:1 york:1 jj:1 useful:1 detailed:1 covered:1 utmost:1 santa:1 concentrated:1 svms:1 http:1 fz:1 zj:1 overly:1 correctly:1 jri:2 shall:2 pz2:7 drawn:2 cb3:1 rcw:1 neither:1 noticing:2 place:2 draw:7 bound:29 hi:4 fold:1 ri:1 bousquet:2 argument:3 min:2 px:1 combination:2 ball:1 llwll:1 jr:1 smaller:1 em:2 son:1 wi:3 appealing:2 den:1 invariant:1 classconditional:1 equation:2 count:1 needed:1 know:3 studying:5 apply:1 observe:2 lxi:1 original:2 rz:2 denotes:1 yx:11 exploit:9 classical:3 quantity:1 font:1 visiting:1 thank:1 mapped:1 neurocolt:2 extent:1 pzm:2 considers:1 length:2 relationship:3 kingdom:1 difficult:1 robert:1 relate:2 unknown:1 maximises:1 upper:1 iti:3 ofb:1 introduced:3 namely:1 california:1 learned:6 able:3 usually:1 xm:2 sparsity:3 explanation:1 ia:7 event:3 indicator:1 smallness:1 scheme:1 ne:1 hm:2 tresp:2 characterises:1 acknowledgement:1 relative:1 loss:9 fully:2 permutation:9 men:4 interesting:1 proven:2 ixl:1 incurred:1 editor:2 pi:2 lo:5 gl:1 supported:1 allow:1 normalised:1 taking:1 face:1 dimension:1 made:1 refinement:1 coincide:1 transaction:1 ml:3 reveals:1 x2m:1 xi:8 why:1 learn:1 mj:5 y2m:1 expansion:1 williamson:4 anthony:2 iiwll:1 main:3 rh:1 whole:1 subsample:3 wmm:2 allowed:1 eural:1 canberra:1 wiley:1 iij:1 xl:3 lie:1 formula:1 luckiness:31 theorem:7 specific:3 pac:1 symbol:1 disregarding:1 virtue:1 serif:1 exists:6 essential:1 vapnik:3 false:1 ilwll:1 importance:2 anu:1 margin:22 cartesian:1 entropy:1 ez:3 mjj:2 hcp:4 g2:2 utilise:1 ma:1 stimulating:1 price:1 generalisation:8 lemma:5 called:2 partly:1 estimable:1 formally:1 support:1 latter:1 ex:1 |
1,214 | 2,106 | Unsupervised Learning of Human Motion
Models
Yang Song, Luis Goncalves, and Pietro Perona
California Institute of Technology, 136-93, Pasadena, CA 9112 5, USA
yangs,luis,perona @vision.caltech.edu
Abstract
This paper presents an unsupervised learning algorithm that can derive
the probabilistic dependence structure of parts of an object (a moving human body in our examples) automatically from unlabeled data. The distinguished part of this work is that it is based on unlabeled data, i.e., the
training features include both useful foreground parts and background
clutter and the correspondence between the parts and detected features
are unknown. We use decomposable triangulated graphs to depict the
probabilistic independence of parts, but the unsupervised technique is
not limited to this type of graph. In the new approach, labeling of the
data (part assignments) is taken as hidden variables and the EM algorithm is applied. A greedy algorithm is developed to select parts and to
search for the optimal structure based on the differential entropy of these
variables. The success of our algorithm is demonstrated by applying it
to generate models of human motion automatically from unlabeled real
image sequences.
1 Introduction
Human motion detection and labeling is a very important but difficult problem in computer
vision. Given a video sequence, we need to assign appropriate labels to the different regions
of the image (labeling) and decide whether a person is in the image (detection). In [8, 7],
a probabilistic approach was proposed by us to solve this problem. To detect and label a
moving human body, a feature detector/tracker (such as corner detector) is first run to obtain
the candidate features from a pair of frames. The combination of features is then selected
based on maximum likelihood by using the joint probability density function formed by the
position and motion of the body. Detection is performed by thresholding the likelihood.
The lower part of Figure 1 depicts the procedure.
One key factor in the method is the probabilistic model of human motion. In order to avoid
exponential combinatorial search, we use conditional independence property of body parts.
In the previous work[8, 7], the independence structures were hand-crafted. In this paper,
we focus on the the previously unresolved problem (upper part of Figure 1): how to learn
the probabilistic independence structure of human motion automatically from unlabeled
training data, meaning that the correspondence between the candidate features and the parts
of the object is unknown. For example when we run a feature detector (such as LucasTomasi-Kanade detector [10]) on real image sequences, the detected features can be from
Unsupervised
Learning
algorithm
Feature
detector/
tracker
Probabilistic
Model of
Human Motion
Unlabeled
Training Data
Feature
detector/
tracker
Detection
and
Labeling
Testing: two frames
Figure 1: Diagram of the system.
Presence of Human?
Localization of parts?
target objects and background clutter with no identity attached to each feature. This case
is interesting because the candidate features can be acquired automatically. Our algorithm
leads to systems able to learn models of human motion completely automatically from real
image sequences - unlabeled training features with clutter and occlusion.
We restrict our attention to triangulated models, since they both account for much correlation between the random variables that represent the position and motion of each body
part, and they yield efficient algorithms. Our goal is to learn the best triangulated model,
i.e., the one that reaches maximum likelihood with respect to the training data. Structure learning has been studied under the graphical model (Bayesian network) framework
([2, 4, 5, 6]). The distinguished part of this paper is that it is an unsupervised learning
method based on unlabeled data, i.e., the training features include both useful foreground
parts and background clutter and the correspondence between the parts and detected features are unknown. Although we work on triangulated models here, the unsupervised technique is not limited to this type of graph.
This paper is organized as follows. In section 2 we summarize the main facts about the
triangulated probability model. In section 3 we address the learning problem when the
training features are labeled, i.e., the parts of the model and the correspondence between
the parts and observed features are known. In section 4 we address the learning problem
when the training features are unlabeled. In section 5 we present some experimental results.
2 Decomposable triangulated graphs
Discovering the probability structure (conditional independence) among variables is important since it makes efficient learning and testing possible, hence some computationally
intractable problems become tractable. Trees are good examples of modeling conditional
(in)dependence [2, 6]. The decomposable triangulated graph is another type of graph which
has been demonstrated to be useful for biological motion detection and labeling [8, 1].
A decomposable triangulated graph [1] is a collection of cliques of size three, where there
is an elimination order of vertices such that when a vertex is deleted, it is only contained
in one triangle and the remaining subgraph is again a collection of triangles until only one
triangle left. Decomposable triangulated graphs are more powerful than trees since each
node can be thought of as having two parents. Similarly to trees, efficient algorithms allow
fast calculation of the maximum likelihood interpretation of a given set of data.
"!,.$-/"!# &0 %("!2')1 %* "!$3)4
+'
Conditional independence among random variables (parts) can be described by a decombe the set of
parts, and
posable triangulated graph. Let
,
, is the measurement for . If the joint probability density function
can be decomposed as a decomposable triangulated graph, it can
"!$%& # ')(* +,(.-/(0
' (01
+ ( 2
- ( 43 ')5+)5-65
')5 2
+ 5 7
-65
9 : <;
>=
@?
C9 :
8
8BA
A
A
where
, 8
9 8 :
C9 : ,
C9 :
A
A are the cliques. 8 8
8
8
8 A
8 A
be written as,
- 0 0 # 0 4 # - 1 # 1
1 & 4 % ' - % 4
0 1 - 0 1 *4
(1)
, and
gives
the elimination order for the decomposable graph.
-/
0 1 4 % %
!0 ! 3
., - 4
,.- 4
,.- 4 ,.- 4 ,.- 4 ,.- 4
., - 4
,.- - 0 4 0 0 4 - 1 1 1 4 - 4 ,.- 4
3 Optimization of the decomposable triangulated graph
FE
are i.i.d samples from a probability density function,
Suppose D
HG
G
G , JI LK , are labeled data. We want to find the
where
MON D is maximized.
MON D is the
decomposable triangulated graph M , such that
probability of graph M being the ?correct? one given the observed data D . Here we use M
to denote both the decomposable graph and the conditional
(in)dependence depicted by the
P
graph. By Bayes? rule, MON D
DHN M
M
D , therefore if we can assume the
priors M are equal for different decompositions, then our goal is to find the structure
triangulated
M which can maximize DHN M C.9 From
: the previous
9 : section, a decomposable
C9 :
A
A , then DHN M
graph M is represented by 8
8
8 A
can be computed as follows,
QSRT VU 1 W YX Z\[ 3 ] !
'(1
+,(
-( Z_[ 3
+ 5 7
- 5
(2)
^
%& ,^
7a
where `
is differential entropy or conditional differential entropy [3] (we consider
continuous random
here). Equation (2) is an approximation which converges
K b variables
c
-4
- 0 0 0 4 - 1 1 1 4 - 4
to equality for
due to the weak Law of Large numbers and definitions and
properties
9 : of differential
C9 : entropy [3,C9 2, 4,
: 5, 6]. We want to find the decomposition
8
8
8BA
A
A such that the above equation can be maximized.
3.1 Greedy search
Though for tree cases, the optimal structure can be obtained efficiently by the maximum
spanning tree algorithm [2, 6], for decomposable triangulated graphs, there is no existing
algorithm which runs in polynomial time and guarantees to the optimal solution [9]. We
develop a greedy algorithm
For
: to grow the graph by the property of decomposable graphs.
9
A
A
each possible choice
of
(the
last
vertex
of
the
last
triangle),
find
the
best
which
? ed 5 ef 5
9 :
A as 8 A , i.e., the
, then
get5 the d best
can maximize `
5 child
? <g
f 5 of edge A
N
vertex (part) that can maximize `
. The next vertex is added one by
one to the existing graph by choosing the best child of all the edges (legal parents)
of the
:
existing graph until all the vertices are added to the graph. For each choice of A , one such
graph can be grown, so there are
candidate graphs. The final result is the graph with the
DFN M among the graphs.
highest hji/k
- 4 /- 4
4
,.- 4
ml - n
? porqts - - l - =u?
rw
v4 o 4 l v44
The above algorithm is efficient. The total search cost is
, which is on the order of
. The algorithm is a greedy algorithm, with no
guarantee that the global optimal solution could be found. Its effectiveness will be explored
through experiments.
- <g ( d ( f ( 4
3.2 Computation of differential entropy - translation invariance
.% %
10 - 4
-/ d ( f ( 4
In thev greedy
and `
,
r= search algorithm, we need to compute `
. If we assume
they are jointly
Gaussian, then the differential entropy can
z|{ that
I
G N }N , where is the dimension and } is the covariance matrix.
be computed by hxiyk
In our applications, position and velocity are used as measurements for each body part, but
humans can be present at different locations of the scene. In order to make the Gaussian
assumption reasonable, translations
s 9 s : s need to be removed. Therefore, we use local coordinate
s
, i.e., we can take one body part (for example 8 ) as
system for each triangle 8
the origin, and use relative positions
body parts. More formally, let denote a
g ( d ( f ( A
g ( d ( for
f ( other
vector of positions
describe
positions
s
? g ( if we
g ( A
d ( ? g ( f ( ? g ( d . ( Then
f ( ?
relative to 8 , we obtain,
.
This can
be written as
, where [12]
- 4
- - 4
, with
4
Z
Z
.
In the greedy search algorithm, the differential entropy of all the possible triplets are
needed and different triplets are with different origins. To reduce computational cost, notice
that
]
[ &
]
[ &
and
3
]
[ &
(4)
!
(5)
G
From the above equations, we can first estimate the mean and covariance } of
(including all the body parts and without removing translation), then take the dimensions
corresponding
9 s
eg ( d ( to the
f ( triangle and use equations (4) and (5) to get the mean and covariance
for
procedure can be applied to pairs (for example,
can be
9 .s Similar
: s
taken as origin for (
)) to achieve translation invariant.
- 4
4 Unsupervised learning of the decomposable graph
0 1
% &%
In this section, we consider the case when only unlabeled data are available. Assume we
K
tI
K
HE
G
have a data set of samples D
. Each sample
,
, is
G
a group of detected features which contains the target object, but
is unlabeled, which
means the correspondence between the candidate features and the parts of the object is
unknown. For example when we run a feature detector (such as Lucas-Tomasi-Kanade
detector [10]) on real image sequences, the detected features can be from target objects and
background clutter with no identity attached to each feature. We want to select the useful
composite parts of the object and learn the probability structure from D .
4.1 All foreground parts observed
Here we first assume that all the foreground parts are observed for each sample. If the labelG
ing for each
is taken as a hidden variable, then the EM algorithm can be used to learn
the probability structure and parameters. Our method was developed from [11], but here
we learn the probabilistic independence structure and all the candidate features are with the
I
FG
FG
same
contains
I type. Let ` G denote the labeling for . If
!
9 features,
9 then ` G is an
M ( M is the back-dimensional vector with each element taken a value from
E
ground clutter label). The observations for the EM algorithm are D
,
` G EG$# , and the parameters to optimize are the probability
the hidden variables are "
(in)dependence structure (i.e. the decomposable triangulated graph) and parameters for its
associated probability density function. We use M to represent both the probability strucG
ture and the parameters. If we assume that
s are independent from each other and ` G
G
only depends on
, then the likelihood function to maximize is,
0
%
0 1
QSRT VU W QSRT VU 1 W & SQ RT W
]
SQ RT ('*] ),+.- '
^ ^ 0/ 1 W 1& Q RT W
&
(6)
'
#
FG
where ` G is the th possible labeling for
, and G is the set of all such labelings. Optimization directly over equation (6) is hard, and
v the EM algorithm solves the optimization
problem iteratively. In EM, for each iteration , we will optimize the function,
W % 1W %
#
#
Q RT
]
]
]
#
0
VU W % 1 U W %
#
QSRT
W % 1
W %
^
&
#
]
1
W % 3 QSRT
0
/
^
& ' ) +.- ' ^
#
]
QSRT
*/ W %
^
^
& ( ' ) +.- ' 0/
#
^
^
0/
W %
(7)
#
G
` G given
where
G is the probability
of ` G
and the decomposable
s
v the observation
probability structure M
. For each iteration ,
G is a fixed number for a hypothesis ` G .
G can be computed as,
#
W %
]
W %
(8)
'*) ^ */
(
#
#
s
HG
We will discuss the computation of ` G
M
below. Under the labeling hypothG
G
` G ,
is divided into the foreground features
, which are parts of the
esis ` G
G
G
object, and background (clutter)
. If the foreground features
are independent of
G
, then,
clutter
W
1 W 2 W
^ 0/
^ */
^ 0/
1 ^ 0/ W 2
1 ^ */ W 2 ^ 0/ 1 W 2 W
(9)
For simplicity, we will assume the priors
` G N M are the same for different ` G , and
M are the same for different graph structures.
If we assume uniform background denG
g G
sities [11, 8], then
, where 8 is the volume of the space a
N` G M
background feature lies in, is the same for different ` G . Under probability decomposition
HG
N ` G M can be computed as in equation (1). Therefore the maximization of
M ,
*/
^
1
*/
#
,.- 4
,.-
W % 7
^
#
,.- # 04
,.- # 4
# 4 - 0 4 3 #
,.-
# 4
*/
equation (7) is equivalent to maximizing,
W % 1 W % C
#
]
X
]
]
& ( '*)
]
& ( '*)
]!
0/ %&
0/
QSRT
1
^
*/
#
W %
QSRT
' 0/( 1
+ */( 7
- *(/ 1& QSRT
+)*/ 5
-60/5
#
#
I
3
(10)
),
For most problems, the number of possible labelings is very large (on the order of
so it is computationally prohibitive to sum over all the possible ` G as in equation (10).
However, if there is one hypothesis labeling `G that is much better than other hypotheses,,
i.e.
G corresponding to `G is much larger than other
G ?s, then
G can be taken as
and other
G ?s as . Hence equation (10) can be approximated as,
# #
W % 1W %
# ##
gG ( dG (
X
#
]
#
fG (
#
#
] ! QSRT
'0/!( 1
+*/"(
-*(/" 1& QSRT
+0/!5
-*/"5
& % &
#
(11)
where
and
are measurements corresponding to the best labeling `#G .
Comparing
with
equation
(2)
and also by the weak law of large numbers, we know for
v
iteration , if the best hypothesis
s ` G is used as the ?true? labeling, then the decomposable
triangulated graph structure M can be obtained through the algorithm described in section
G
3. One approximation we make here is that the best hypothesis labeling ` G for each
is
really dominant among all the possible labelings so that hard assignment for labelings can
be used. This is similar to the situation of K-means vs. mixture of Gaussian for clustering
problems. We evaluate this approximation in experiments.
#
#
The whole algorithm can be summarized as follows. Given some random initial
v v guess of
the decomposable graph structure M and its parameters, then for iteration , ( is from
until the algorithm converges),
s
FG
E step: for each
, use M
to find the best labeling `G and then compute the differential
entropies;
M step: use the differential
entropies to run the greedy graph growing algorithm described
s
in section 3 and get M .
0
#
4.2 Dealing with missing parts (occlusion)
So far we assume that all the parts are observed. In the case of some parts missing, the
measurements for the missing parts can be taken as additional hidden variables [11], and
the EM algorithm can be modified to handle the missing parts.
For each hypothesis ` G , let G denote the measurements
of the observed parts, G be
A
G
G
G A A be the measurements of
the measurements for the missing parts, and
the whole object (to reduce clutter in the notation, we assume that the dimensions can be
sorted in this way). For each EM iteration,
we need to compute G
and } G
to obtain
s
the differential entropies and then M with its parameters. Taking ` G and G as hidden
variables, we can get,
2
[
]
Z
2
2
Z
[
]
2
]
! [
(12)
!
Z
.
!
2
(13)
! ! , and
!
.
! !
! ! .
G
All the expectations
are conditional
expectations with respect to
` G
` G and
s
G are the measurements of the observed
decomposable graph structure M
. Therefore,
s
foreground parts under ` G
`G . Since M
is Gaussian distributed, conditional expec G and G s G A can be computed from observed parts G and the mean
tation
Where
- 4
-4
!
a
0
#
- 04
M
and covariance matrix of
0
#
.
5 Experiments
We tested the greedy algorithm on labeled motion capture data (Johansson displays) as in
[8], and the EM-like algorithm on unlabeled detected features from real image sequences.
5.1 Motion capture data
Our motion capture data consist of the 3-D positions of 14 markers fixed rigidly on a
subject?s body. These positions were tracked with 1mm accuracy as the subject walked
back and forth, and projected to 2-D.
Under Gaussian assumption, we first estimated the joint probability density function (mean
and covariance) of the data. From the estimated mean and covariance, we can compute
differential entropies for all the possible triplets and pairs and further run the greedy search
algorithm to find the approximated best triangulated model. Figure 2(a) shows the expected
likelihood (differential entropy) of the estimated joint pdf, of the best triangulated model
from the greedy algorithm, of the hand-constructed model from [8], and of randomly generated models. The greedy model is clearly superior to the hand-constructed model and the
random models. The gap to the original joint pdf is partly due to the strong conditional independence assumptions of the triangulated model, which are an approximation of the true
data?s pdf. Figure 2(b) shows the expected likelihood using 50 synthetic datasets. Since
these datasets were generated from 50 triangulated models, the greedy algorithm (solid
curve) can match the true model (dashed curve) extremely well. The solid line with error
bars are the expected likelihoods of random triangulated models.
?110
?135
?140
expected log likelihood
expected log likelihood
estimated joint pdf
?120
best trangulated model from greedy search
?130
?145
triangulated model used in previous papers
?150
?140
?155
?150
?160
randomly generated triangulated models
?160
0
500
1000
1500
2000
2500
3000
index of randomly generated triangulated models
?165
0
10
20
30
40
50
60
index of randomly generated triangulated models
(a)
(b)
Figure 2: Evaluation of greedy search.
5.2 Real image sequences
There are three types of sequences used here: (I) a subject walks from left to right (Figure
3(a,b)); (II) a subject walks from right to left; (III) a subject rides a bike from left to right
(Figure3(c,d)). Left-to-right walking motion models were learned from type I sequences
and tested on all types of sequences to see if the learned model can detect left-to-right
walking and label the body parts correctly. The candidate features were obtained from a
Lucas-Tomasi-Kanade algorithm [10] on two frames. We used two frames to simulate the
difficult situation, where due to extreme body motion or to loose and textured clothing and
occlusion, tracking is extremely unreliable and each feature?s lifetime is short.
Evaluation of the EM-like algorithm. As described in section 4.1, one approximation
we made is taking the best hypothesis labeling instead of summing over all the possible
hypotheses (equation (11)). This approximation was evaluated by checking how the loglikelihoods evolve with EM iterations and if they converge. Figure 4(a) shows the results of
learning a 12-feature model. We used random initializations, and each curve of Figure 4(a)
corresponds to one such random initialization. From Figure 4(a) we can see that generally
the log-likelihoods grow and converge well with the iterations of EM.
Models obtained. Figure 4 (b) and (c) show the best model obtained after we ran the
EM algorithms for 11 times. Figure 4(b) gives the mean positions and mean velocities
(shown in arrows) of the parts. Figure 4(c) shows the learned decomposable triangulated
probabilistic structure. The letter labels show the body parts correspondence.
Figure 3 shows samples of the results. The red dots (with letter labels) are the maximum
likelihood configuration from the left-to-right walking model. The horizontal bar at the
bottom left of each frame shows the likelihood
of the
The short vers
? best
s configuration.
tical bar gives the threshold where
G for all the test data. If
* ,
,
G
B
F
J CD
#
G
F
J C
B
D
H
H
H
H
E
IE A L
I
(a)walking detected
K
A
L
(b)detected
I
F G B
C D
E A
L
(c)not detected
F
C
G
I
B
DE
A
L
(d)not detected
Figure 3: Sample frames from left-to-right walking (a-b) and biking sequences (c-d). The dots
(either filled or empty) are the features selected by Tomasi-Kanade algorithm [10] on two frames.
The filled dots (with letter labels) are the maximum likelihood configuration from the left-to-right
walking model. The horizontal bar at the bottom left of each frame shows the likelihood of the best
configuration. The short vertical bar gives the threshold for detection.
?100
G
40
log likelihood
?120
J
C
G
B
C
D
80
?140
100
?160
H
H
120
?180
140
160
?200
180
?220
0
F
B
D
F
J
60
5
10
iterations of EM
15
20
200
I
100
E
150
K A L
200
I
250
E
K
A
L
(a)
(b)
(c)
Figure 4: (a) Evaluation of the EM-like algorithm. Log-likelihood vs. iterations of EM for different
random initializations. (b) and (c) show the best model obtained after we ran the EM-like algorithms
for 11 times.
the likelihood is greater than the threshold, a left-to-right walking person is detected.The
detection rate is 100% for the left-to-right walking vs. right-to-left walking, and 87% for
the left-to-right walking vs. left-to-right biking.
6 Conclusions
In this paper we have described a method for learning the structure and parameters of a
decomposable triangulated graph in an unsupervised fashion from unlabeled data. We have
applied this method to learn models of biological motion that can be used to reliably detect
and label biological motion.
Acknowledgments
Funded by the NSF Engineering Research Center for Neuromorphic Systems Engineering
(CNSE) at Caltech (NSF9402726), and by an NSF National Young Investigator Award to
PP (NSF9457618). We thank Charless Fowlkes for bringing the Chow and Liu?s paper to
our attention. We thank Xiaolin Feng for providing the real image sequences.
References
[1] Y. Amit and A. Kong, ?Graphical templates for model registration?, IEEE Transactions on
Pattern Analysis and Machine Intelligence, 18:225?236, 1996.
[2] C.K. Chow and C.N. Liu, ?Approximating discrete probability distributions with dependence
trees?, IEEE Transactions on Information Theory, 14:462?467, 1968.
[3] T.M. Cover and J.A. Thomas, Elements of Information Theory, John Wiley and Sons, 1991.
[4] N. Friedman and M. Goldszmidt, ?Learning bayesian networks from data?, Technical report,
AAAI 1998 Tutorial, http://robotics.stanford.edu/people/nir/tutorial/, 1998.
[5] M.I. Jordan, editor, Learning in Graphical Models, MIT Press, 1999.
[6] M. Meila and M.I. Jordan, ?Learning with mixtures of trees?, Journal of Machine Learning
Rearch, 1:1?48, 2000.
[7] Y. Song, X. Feng, and P. Perona, ?Towards detection of human motion?, In Proc. IEEE CVPR
2000, volume 1, pages 810?817, June 2000.
[8] Y. Song, L. Goncalves, E. Di Bernardo, and P. Perona, ?Monocular perception of biological
motion in johansson displays?, Computer Vision and Image Understanding, 81:303?327, 2001.
[9] Nathan Srebro, ?Maximum likelihood bounded tree-width markov networks?, In UAI, pages
504?511, San Francisco, CA, 2001.
[10] C. Tomasi and T. Kanade, ?Detection and tracking of point features?, Tech. Rep. CMU-CS-91132,Carnegie Mellon University, 1991.
[11] M. Weber, M. Welling, and P. Perona, ?Unsupervised learning of models for recognition?, In
Proc. ECCV, volume 1, pages 18?32, June/July 2000.
[12] Markus Weber, Unsupervised Learning of Models for Object Recognition, Ph.d. thesis, Caltech,
May 2000.
| 2106 |@word kong:1 polynomial:1 johansson:2 decomposition:3 covariance:6 solid:2 initial:1 configuration:4 contains:2 liu:2 existing:3 comparing:1 written:2 luis:2 john:1 depict:1 v:4 greedy:14 selected:2 discovering:1 prohibitive:1 guess:1 intelligence:1 short:3 node:1 location:1 constructed:2 differential:12 become:1 acquired:1 expected:5 growing:1 decomposed:1 automatically:5 notation:1 bounded:1 bike:1 developed:2 guarantee:2 ti:1 bernardo:1 dfn:1 engineering:2 local:1 tation:1 rigidly:1 initialization:3 studied:1 limited:2 acknowledgment:1 testing:2 vu:4 sq:2 procedure:2 thought:1 composite:1 get:3 unlabeled:12 applying:1 optimize:2 equivalent:1 demonstrated:2 missing:5 maximizing:1 center:1 attention:2 decomposable:21 simplicity:1 rule:1 handle:1 coordinate:1 biking:2 target:3 suppose:1 hypothesis:8 origin:3 velocity:2 element:2 approximated:2 recognition:2 walking:10 labeled:3 xiaolin:1 observed:8 bottom:2 capture:3 region:1 highest:1 removed:1 ran:2 localization:1 completely:1 triangle:6 textured:1 joint:6 represented:1 grown:1 fast:1 describe:1 detected:11 labeling:14 choosing:1 mon:3 larger:1 solve:1 stanford:1 cvpr:1 jointly:1 final:1 sequence:12 unresolved:1 loglikelihoods:1 subgraph:1 achieve:1 forth:1 parent:2 empty:1 converges:2 object:10 derive:1 develop:1 strong:1 solves:1 c:1 triangulated:28 correct:1 human:12 elimination:2 assign:1 really:1 biological:4 clothing:1 mm:1 tracker:3 ground:1 proc:2 label:8 combinatorial:1 mit:1 clearly:1 gaussian:5 modified:1 avoid:1 focus:1 june:2 likelihood:19 tech:1 detect:3 chow:2 perona:5 pasadena:1 hidden:5 labelings:4 among:4 lucas:2 equal:1 having:1 unsupervised:10 foreground:7 report:1 randomly:4 dg:1 national:1 occlusion:3 friedman:1 detection:9 evaluation:3 mixture:2 extreme:1 hg:3 tical:1 edge:2 tree:8 filled:2 walk:2 modeling:1 cover:1 assignment:2 maximization:1 neuromorphic:1 cost:2 vertex:6 uniform:1 synthetic:1 person:2 density:5 ie:1 probabilistic:8 v4:1 again:1 aaai:1 thesis:1 corner:1 account:1 de:1 summarized:1 depends:1 performed:1 red:1 bayes:1 walked:1 formed:1 accuracy:1 efficiently:1 maximized:2 yield:1 weak:2 bayesian:2 detector:8 reach:1 ed:1 definition:1 pp:1 associated:1 di:1 organized:1 back:2 evaluated:1 though:1 lifetime:1 correlation:1 until:3 hand:3 horizontal:2 marker:1 usa:1 true:3 hence:2 equality:1 iteratively:1 eg:2 width:1 gg:1 pdf:4 motion:19 image:10 meaning:1 weber:2 ef:1 charles:1 superior:1 ji:1 tracked:1 attached:2 volume:3 interpretation:1 he:1 measurement:8 mellon:1 meila:1 similarly:1 dot:3 funded:1 moving:2 ride:1 dominant:1 rep:1 success:1 caltech:3 additional:1 greater:1 deng:1 converge:2 maximize:4 dashed:1 ii:1 july:1 ing:1 technical:1 match:1 calculation:1 divided:1 award:1 vision:3 expectation:2 cmu:1 iteration:9 represent:2 robotics:1 background:7 want:3 diagram:1 grow:2 bringing:1 subject:5 effectiveness:1 jordan:2 yang:2 presence:1 iii:1 ture:1 independence:8 restrict:1 reduce:2 whether:1 song:3 dhn:3 useful:4 generally:1 clutter:9 ph:1 rw:1 generate:1 http:1 nsf:2 tutorial:2 notice:1 estimated:4 correctly:1 hji:1 discrete:1 carnegie:1 group:1 key:1 threshold:3 deleted:1 registration:1 graph:35 pietro:1 sum:1 run:6 letter:3 powerful:1 reasonable:1 decide:1 display:2 correspondence:6 scene:1 markus:1 nathan:1 simulate:1 extremely:2 combination:1 em:16 son:1 invariant:1 taken:6 computationally:2 equation:11 legal:1 previously:1 monocular:1 discus:1 loose:1 needed:1 know:1 tractable:1 available:1 appropriate:1 c9:6 distinguished:2 fowlkes:1 original:1 thomas:1 remaining:1 include:2 clustering:1 graphical:3 qsr:11 yx:1 amit:1 approximating:1 feng:2 added:2 dependence:5 thank:2 figure3:1 spanning:1 index:2 providing:1 difficult:2 fe:1 ba:2 reliably:1 unknown:4 upper:1 vertical:1 observation:2 datasets:2 markov:1 situation:2 frame:8 pair:3 tomasi:4 california:1 learned:3 address:2 able:1 bar:5 below:1 pattern:1 perception:1 summarize:1 including:1 video:1 technology:1 lk:1 nir:1 prior:2 understanding:1 checking:1 evolve:1 relative:2 law:2 goncalves:2 interesting:1 srebro:1 thresholding:1 editor:1 cd:1 translation:4 eccv:1 last:2 allow:1 institute:1 template:1 taking:2 fg:5 distributed:1 curve:3 dimension:3 collection:2 made:1 projected:1 san:1 far:1 welling:1 transaction:2 unreliable:1 clique:2 ml:1 global:1 dealing:1 uai:1 vers:1 summing:1 francisco:1 search:9 continuous:1 triplet:3 kanade:5 learn:7 ca:2 z_:1 main:1 arrow:1 whole:2 child:2 body:13 crafted:1 depicts:1 fashion:1 wiley:1 position:9 exponential:1 candidate:7 lie:1 young:1 removing:1 explored:1 expec:1 intractable:1 consist:1 esis:1 gap:1 entropy:12 depicted:1 contained:1 tracking:2 corresponds:1 conditional:9 identity:2 goal:2 sorted:1 towards:1 hard:2 total:1 invariance:1 experimental:1 partly:1 select:2 formally:1 people:1 goldszmidt:1 investigator:1 evaluate:1 tested:2 |
1,215 | 2,107 | Tree-based reparameterization for
approximate inference on loopy graphs
Martin J. Wainwright, Tommi Jaakkola, and Alan S. Will sky
Department of Electrical Engineering and Computer Science
Massachusetts Institute of Technology
Cambridge, MA 02139
[email protected]
[email protected]
[email protected]
Abstract
We develop a tree-based reparameterization framework that provides a new conceptual view of a large class of iterative algorithms
for computing approximate marginals in graphs with cycles. It
includes belief propagation (BP), which can be reformulated as a
very local form of reparameterization. More generally, we consider
algorithms that perform exact computations over spanning trees
of the full graph. On the practical side, we find that such tree
reparameterization (TRP) algorithms have convergence properties
superior to BP. The reparameterization perspective also provides
a number of theoretical insights into approximate inference, including a new characterization of fixed points; and an invariance
intrinsic to TRP /BP. These two properties enable us to analyze
and bound the error between the TRP /BP approximations and
the actual marginals. While our results arise naturally from the
TRP perspective, most of them apply in an algorithm-independent
manner to any local minimum of the Bethe free energy. Our results also have natural extensions to more structured approximations [e.g. , 1, 2].
1
Introduction
Given a graphical model, one important problem is the computation of marginal
distributions of variables at each node. Although highly efficient algorithms exist
for this task on trees, exact solutions are prohibitively complex for more general
graphs of any substantial size. This difficulty motivates the use of approximate
inference algorithms, of which one of the best-known and most widely studied is
belief propagation [3], also known as the sum-product algorithm in coding [e.g., 4].
Recent work has yielded some insight into belief propagation (BP). Several researchers [e.g., 5, 6] have analyzed the single loop case, where BP can be reformulated as a matrix powering method. For Gaussian processes on arbitrary graphs,
two groups [7, 8] have shown that the means are exact when BP converges. For
graphs corresponding to turbo codes, Richardson [9] established the existence of
fixed points, and gave conditions for their stability. More recently, Yedidia et al. [1]
showed that BP corresponds to constrained minimization of the Bethe free energy,
and proposed extensions based on Kikuchi expansions [10]. Related extensions to
BP were proposed in [2]. The paper [1] has inspired other researchers [e.g., 11, 12] to
develop more sophisticated algorithms for minimizing the Bethe free energy. These
advances notwithstanding, much remains to be understood about the behavior of
BP.
The framework of this paper provides a new conceptual view of various algorithms
for approximate inference, including BP. The basic idea is to seek a reparameterization of the distribution that yields factors which correspond, either exactly or
approximately, to the desired marginal distributions. If the graph is acyclic (i.e.,
a tree) , then there exists a unique reparameterization specified by exact marginal
distributions over cliques. For a graph with cycles, we consider the idea of iteratively reparameterizing different parts of the distribution, each corresponding to an
acyclic subgraph. As we will show, BP can be interpreted in exactly this manner ,
in which each reparameterization takes place over a pair of neighboring nodes. One
of the consequences of this interpretation is a more storage-efficient "message-free"
implementation of BP. More significantly, this interpretation leads to more general
updates in which reparameterization is performed over arbitrary acyclic subgraphs,
which we refer to as tree-based reparameterization (TRP) algorithms.
At a low level, the more global TRP updates can be viewed as a tree-based schedule
for message-passing. Indeed, a practical contribution of this paper is to demonstrate that TRP updates tend to have better convergence properties than local
BP updates. At a more abstract level, the reparameterization perspective provides
valuable conceptual insight, including a simple tree-consistency characterization of
fixed points, as well as an invariance intrinsic to TRP /BP. These properties allow
us to derive an exact expression for the error between the TRP /BP approximations
and the actual marginals. Based on this exact expression, we derive computable
bounds on the error. Most of these results, though they emerge very naturally in
the TRP framework , apply in an algorithm-independent manner to any constrained
local minimum of the Bethe free energy, whether obtained by TRP /BP or an alternative method [e.g. , 11, 12]. More details of our work can be found in [13, 14].
1.1
Basic notation
An undirected graph Q = (V, ?) consists of a set of nodes or vertices V = {l , ... ,N}
that are joined by a set of edges ?. Lying at each node s E V is a discrete
random variable Xs E {a, ... ,m - I}. The underlying sample space X N is the
set of all N vectors x = {x s I S E V} over m symbols, so that IXNI = m N .
We focus on stochastic processes that are Markov with respect to Q, so that the
Hammersley-Clifford theorem [ e.g., 3] guarantees that the distribution factorizes
as p(x) ex: [lc Ee 'l/Jc(xc) where 'l/Jc(xc) is a compatibility function depending only
on the subvector Xc = {xs I SEC} of nodes in a particular clique C. Note that
each individual node forms a singleton clique, so that some of the factors 'l/Jc may
involve functions of each individual variable. As a consequence, if we have independent measurements Ys of Xs at some (or all) of the nodes, then Bayes' rule
implies that the effect of including these measurements - i.e., the transformation
from the prior distribution p(x) to the conditional distribution p(x I y) - is simply
to modify the singleton factors. As a result, throughout this paper, we suppress
explicit mention of measurements, since the problem of computing marginals for
either p(x) or p(x Iy) are of identical structure and complexity. The analysis of
this paper is restricted to graphs with singleton ('l/Js) and pairwise ('l/Jst} cliques.
However, it is straightforward to extend reparameterization to larger cliques, as in
cluster variational methods [e.g., 10].
1.2
Exact tree inference as reparameterization
Algorithms for optimal inference on trees have appeared in the literature of various fields [e.g., 4, 3]. One important consequence of the junction tree representation [15] is that any exact algorithm for optimal inference on trees actually computes
marginal distributions for pairs (s, t) of neighboring nodes. In doing so, it produces
an alternative factorization p(x) = TI sEV P s TI(s,t)E? Pst/(PsPt ) where Ps and Pst
are the single-node and pairwise marginals respectively. This {Ps , Pst} representation can be deduced from a more general factorization result on junction trees [e.g.
15]. Thus, exact inference on trees can be viewed as computing a reparameterized factorization of the distribution p(x) that explicitly exposes the local marginal
distributions.
2
Tree-based reparameterization for graphs with cycles
The basic idea of a TRP algorithm is to perform successive reparameterization updates on trees embedded within the original graph. Although such updates are
applicable to arbitrary acyclic substructures, here we focus on a set T 1 , ... , TL
of embedded spanning trees. To describe TRP updates, let T be a pseudomarginal probability vector consisting of single-node marginals Ts(x s ) for 8 E V;
and pairwise joint distributions Tst (x s, Xt) for edges (s, t) E [. Aside from positivity and normalization (Lx s Ts = 1; L xs , xt Tst = 1) constraints, a given vector T is arbitraryl , and gives rises to a parameterization of the distribution as
p(x; T) ex: TI sEV Ts TI(S,t)E? Tst/ {(L x. Tst)(L Xt Tst )}, where the dependence of Ts
and Tst on x is omitted for notational simplicity. Ultimately, we shall seek vectors
T that are consistent - i.e. , that belong to <C = {T I Lx. Tst = Tt \;/ (8, t) E [}. In
the context of TRP, such consistent vectors represent approximations to the exact
marginals of the distribution defined by the graph with cycles.
We shall express TRP as a sequence of functional updates Tn I-t T n+1 , where
superscript n denotes iteration number. We initialize at TO via T~t = Ii 'l/Js'I/Jt'I/Jst
and T~ = Ii 'l/Js TItEN(S) [L X t 'l/Jst'I/Jt], where Ii denotes a normalization factor; and
N(8) is the set of neighbors of node 8. At iteration n, we choose some spanning
tree Ti(n) with edge set [i(n), and factor the distribution p(x; Tn) into a product
of two terms
ex:
(la)
ex:
(lb)
corresponding, respectively, to terms in the spanning tree; and residual terms over
edges in [/ [i(n) removed to form Ti(n). We then perform a reparameterization
update on pi(n) (x; Tn) - explicitly:
pi(n)
(x'; Tn)
for all (s,t) E
[i(n)
(2)
x, s.t( x ~ ,x;)=(x. ,xtl
with a similar update for the single-node marginals {Ts I s E V}. These marginal
computations can be performed efficiently by any exact tree algorithm applied to
Ti(n). Elements of T n+1 corresponding to terms in ri(n) (x; Tn) are left unchanged
lIn general, T need not be the actual marginals for any distribution.
(i.e., Ts~+l = Tst for all (8, t) E E/Ei(n)) . The only restriction placed on the spanning
tree set T 1, ... ,TL is that each edge (8, t) E E belong to at least one spanning tree.
For practical reasons, it is desirable to choose a set of spanning trees that leads to
rapid mixing throughout the graph. A natural choice for the spanning tree index
i(n) is the cyclic ordering, in which i(n) == n(modL) + 1.
2.1
BP as local reparameterization
Interestingly, BP can be reformulated in a "message-free" manner as a sequence
of local rather than global reparameterization operations. This message-free version of BP directly updates approximate marginals, Ts and Tst, with initial values determined from the initial messages M~t and the original compatibility functions of the graphical model as T~ = Ii 'l/Js ITu EN(S) M~s for all 8 E V and
T~t = Ii 'l/Jst'l/Js'l/Jt ITu EN(s)/t M~s ITuEN(t) /s M~t for all (8, t) E E, where Ii denotes a normalization factor. At iteration n, these quantities are updated according
to the following recursions:
(3a)
T;'t
(3b)
The reparameterization form of BP decomposes the graph into a set of two-node
trees (one for each edge (8, t)); performs exact inference on such tree via equation (3b); and merges the marginals from each tree via equation (3a). It can be
shown by induction [see 13] that this simple reparameterization algorithm is equivalent to the message-passing version of BP.
2.2
Practical advantages of TRP updates
Since a single TRP update suffices to transmit information globally throughout the
graph, it might be expected to have better convergence properties than the purely
local BP updates. Indeed, this has proven to be the case in various experiments that
we have performed on two graphs (a single loop of 15 nodes, and a 7 x 7 grid). We
find that TRP tends to converge 2 to 3 times faster than BP on average (rescaled
for equivalent computational cost); more importantly, TRP will converge for many
problems where BP fails [13]. Further research needs to address the optimal choice
of trees (not necessarily spanning) in implementing TRP.
3
Theoretical results
The TRP perspective leads to a number of theoretical insights into approximate
inference, including a new characterization of fixed points, an invariance property,
and error analysis.
3.1
Analysis of TRP updates
Our analysis of TRP updates uses a cost function that is an approximation to the
Kullback-Leibler divergence between p(x; T) and p(x; U) - namely, the quantity
Xs
Given an arbitrary U E C, we show that successive iterates {Tn} of TRP updates
satisfy the following "Pythagorean" identity:
G(U ; T n)
=
G(U ; T n+ l ) + G(T n+1; T n)
(4)
which can be used to show that TRP fixed points T * satisfy the necessary conditions
to be local minima of G subject to the constraint T * E C. The cost function G,
though distinct from the Bethe free energy [1] , coincides with it on the constraint
set C, thereby allowing us to establish the equivalence of TRP and BP fixed points.
3.2
Characterization of fixed points
From the reparameterization perspective arises an intuitive characterization of any
TRP /BP fixed point T *. Shown in Figure l(a) is a distribution on a graph with
T 1:
T~T;
T4 ;
T3~
T2~
T; T;
T; T ~
T5:
T 1:
TtT;
T2~
T2*T;
T3~
T; T ~
(a) Fixed point on full graph
(b) Tree consistency condition.
Figure 1. Illustration of fixed point consistency condition. (a) Fixed point T * =
{T;, T;t } on the full graph with cycles. (b) Illustration of consistency condition on
an embedded tree. The quantities {T;, T;t } must be exact marginal probabilities
for any tree embedded within the full graph.
cycles, parameterized according to the fixed point T * = {Ts*t, T;}. The consistency
condition implies that if edges are removed from the full graph to form a spanning
tree, as shown in panel (b) , then the quantities Ts*t and Ts* correspond to exact
marginal distributions over the tree. This statement holds for any acyclic substructure embedded within the full graph with cycles - not just the spanning trees
Tl , ... ,TL used to implement TRP. Thus, algorithms such as TRP /BP attempt
to reparameterize a distribution on a graph with cycles so that it is consistent with
respect to each embedded tree.
It is remarkable that the existence of such a parameterization (though obvious for
trees) should hold for a positive distribution on an arbitrary graph. Also noteworthy
is the parallel to the characterization of max-product 2 fixed points obtained by
Freeman and Weiss [16]. Finally, it can be shown [13, 14] that this characterization,
though it emerged very naturally from the TRP perspective, applies more generally
to any constrained local minimum of the Bethe free energy, whether obtained by
TRP /BP, or an alternative technique [e.g., 11, 12].
2Max-product is a related but different algorithm for computing approximate MAP
assignments in graphs with cycles.
3.3
Invariance of the distribution
A fundamental property of TRP updates is that they leave invariant the full distribution on the graph with cycles. This invariance follows from the decomposition of
equation (1): in particular, the distribution pi(n) (x; Tn) is left invariant by reparameterization; and TRP does not change terms in ri(n) (x; Tn). As a consequence, the
overall distribution remains invariant - i.e., p(x; Tn) == p(x; TO) for all n. By continuity of the map T f-7 p(x; T) , it follows that any fixed point T* of the algorithm
also satisfies p(x; T*) == p(x; TO). This fixed point invariance is also an algorithmindependent result - in particular, all constrained local minima of the Bethe free
energy, regardless of how they are obtained, are invariant in this manner [13, 14].
This invariance has a number of important consequences. For example, it places
severe restrictions on cases (other than trees) in which TRP /BP can be exact;
see [14] for examples. In application to the linear-Gaussian problem, it leads to an
elementary proof of a known result [7, 8] - namely, the means must be exact if the
BP updates converge.
3.4
Error analysis
Lastly, we can analyze the error arising from any TRP /BP fixed point T* on an
arbitrary graph. Of interest are the exact single-node marginals Ps of the original distribution p(x; TO) defined by the graph with cycles, which by invariance are
equivalent to those of p(x; T*). Now the quantities Ts* have two distinct interpretations: (a) as the TRP /BP approximations to the actual single-node marginals on
the full graph; and (b) as the exact marginals on any embedded tree (as in Figure 1).
This implies that the approximations T; are related to the actual marginals P s on
the full graph by a relatively simple perturbation - namely, removing edges from
the full graph to reveal an embedded tree. From this observation, we can derive the
following exact expression for the difference between the actual marginal PS;j and
the TRP /BP approximation 3 T;j:
ri(X; T * ) }
.J
lEpi (x;T * ) [{ Z(T*) - 1 J(x s = J)
(5)
where i E {1, ... ,L} is an arbitrary spanning tree index; pi and ri are defined in
equation (1a) and (1b) respectively; Z(T*) is the partition function of p(x; T*);
J(xs = j) is an indicator function for Xs to take the value j; and lEpi (x;T * ) denotes
expectation using the distribution pi(x; T*).
Unfortunately, while the tree distribution pi (x; T*) is tractable, the argument of the
expectation includes all terms r i (x ; T*) removed from the original graph to form
spanning tree Ti. Moreover, computing the partition function Z (T*) is intractable.
These difficulties motivate the development of bounds on the error.
In [14], we use convexity arguments to derive a particular set of bounds on the
approximation error. Such error bounds, in turn, can be used to compute upper
and lower bounds on the actual marginals Ps;l. Figure 2 illustrates the TRP /BP
approximation, as well as these bounds on the actual marginals for a binary process
on a 3 x 3 grid under two conditions. Note that the tightness of the bounds is closely
related to approximation accuracy. Although it is unlikely that these bounds will
remain quantitatively useful for general problems on large graphs, they may still
yield useful qualitative information.
3The notation T;;j denotes the
/h
element of the vector T; .
Bounds on single node marginals
Bounds on single node marginals
0.9
0.9
0.8
0.8
0.7
:;:::'0.6
"
:5-"b.5
e
"- o.
0.2
0.1
?1~~--~--~4~~5---6~~~~~~
Node number
(a) Weak potentials
4
5
6
Node number
(b) Strong mixed potentials
Figure 2. Behavior of bounds on 3 x 3 grid. Plotted are the actual marginals P s;l
versus the TRP approximations T;'l> as well as upper and lower bounds on the
actual marginals. (a) For weak potentials, TRP /BP approximation is excellent;
bounds on exact marginals are tight. (b) For strong mixed potentials, approximation is poor. Bounds are looser, and for certain nodes, the TRP /BP approximation
lies above the upper bounds on the actual marginal P 8 ;1 .
Much of the analysis of this paper -- including reparameterization, invariance,
and error analysis -- can be extended [see 14] to more structured approximation
algorithms [e.g., 1, 2]. Figure 3 illustrates the use of bounds in assessing when to
use a more structured approximation. For strong attractive potentials on the 3 x 3
grid, the TRP /BP approximation in panel (a) is very poor, as reflected by relatively
loose bounds on the actual marginals. In contrast, the Kikuchi approximation in
(b) is excellent, as revealed by the tightness of the bounds.
4
Discussion
The TRP framework of this paper provides a new view of approximate inference;
and makes both practical and conceptual contributions. On the practical side, we
find that more global TRP updates tend to have better convergence properties than
local BP updates. The freedom in tree choice leads to open problems of a graphtheoretic nature: e.g., how to choose trees so as to guarantee convergence, or to
optimize the rate of convergence?
Among the conceptual insights provided by the reparameterization perspective are
a new characterization of fixed points; an intrinsic invariance; and analysis of the
approximation error. Importantly, most of these results apply to any constrained
local minimum of the Bethe free energy, and have natural extensions [see 14] to
more structured approximations [e.g., 1, 2].
Acknowledgments
This work partially funded by ODDR&E MURI Grant DAAD19-00-1-0466; by ONR
Grant N00014-00-1-0089; and by AFOSR Grant F49620-00-1-0362; MJW also supported
by NSERC 1967 fellowship.
References
[1] J. Yedidia, W. T. Freeman, and Y. Weiss. Generalized belief propagation. In NIPS
13, pages 689- 695. MIT Press, 2001.
Bounds on single node marginals
-
- - - -0 - - - -0- -
Bounds on single node marginals
-
e- - - -
M
M
0.8
_ 0 - - - - 0 - - - -0- - - - ?l - - - -
;o: V
"
:5-"b.
"
:5-"b.5
?>
a.. 0.4
?>
a.. 0.4
0.3
0.3
:~ II=-:= ~~r~~lured
0.2 rr -+-:Ac-,tu--:
al----.
0.1
-+- 0 ?
TAP I BP
Bounds
?1~~==~~-4~~5~~-~-~~
Node number
(a) TRP /BP
approx. 1
~rl=-e~B=o=un=ds~==~~~~-~-~-~
?1
4
5
Node number
(b) Kikuchi
Figure 3. When to use a more structured approximation? (a) For strong attractive potentials on the 3 x 3 grid, BP approximation is poor, as reflected by loose
bounds on the actual marginal. (b) Kikuchi approximation [1] for same problem
is excellent; corresponding bounds are tight.
[2] T. P. Minka. A family of algorithms for approximate Bayesian inference. PhD thesis,
MIT Media Lab, 2001.
[3] J. Pearl. Probabilistic reasoning in intelligent systems. Morgan Kaufman, San Mateo,
1988.
[4] F. Kschischang and B. Frey. Iterative decoding of compound codes by probability
propagation in graphical models. IEEE Sel. Areas Comm., 16(2):219- 230, February
1998.
[5] J. B. Anderson and S. M. Hladnik. Tailbiting map decoders. IEEE Sel. Areas Comm.,
16:297- 302, February 1998.
[6] Y. Weiss. Correctness of local probability propagation in graphical models with loops.
Neural Computation, 12:1-41, 2000.
[7] Y. Weiss and W. T. Freeman. Correctness of belief propagation in Gaussian graphical
models of arbitrary topology. In NIPS 12, pages 673- 679 . MIT Press, 2000 .
[8] P. Rusmevichientong and B. Van Roy. An analysis of turbo decoding with Gaussian
densities. In NIPS 12, pages 575- 581. MIT Press, 2000.
[9] T. Richardson. The geometry of turbo-decoding dynamics. IEEE Trans. Info. Theory,
46(1):9- 23, January 2000.
[10] R. Kikuchi. The theory of cooperative phenomena. Physical Review, 81:988- 1003,
1951.
[11] M. Welling and Y. Teh. Belief optimization: A stable alternative to loopy belief
propagation. In Uncertainty in Artificial Intelligence, July 2001.
[12] A. Yuille. A double-loop algorithm to minimize the Bethe and Kikuchi free energies.
Neural Computation, To appear, 2001.
[13] M. J . Wainwright, T. Jaakkola, and A. S. Willsky. Tree-based reparameterization for
approximate estimation on graphs with cycles. LIDS Tech. report P-2510: available
at http://ssg.rnit.edu/group/rnjyain/rnjyain.shtrnl, May 2001.
[14] M. Wainwright . Stochastic processes on graphs with cycles: geometric and variational
approaches. PhD thesis, MIT, Laboratory for Information and Decision Systems,
January 2002.
[1 5] S. L. Lauritzen. Graphical models. Oxford University Press, Oxford, 1996.
[16] W. Freeman and Y. Weiss. On the optimality of solutions of the max-product belief
propagation algorithm in arbitrary graphs. IEEE Trans. Info. Theory, 47:736- 744,
2001.
| 2107 |@word version:2 open:1 seek:2 decomposition:1 mention:1 thereby:1 initial:2 cyclic:1 interestingly:1 must:2 partition:2 update:21 aside:1 intelligence:1 parameterization:2 provides:5 characterization:8 iterates:1 node:25 successive:2 lx:2 qualitative:1 consists:1 xtl:1 manner:5 pairwise:3 expected:1 indeed:2 rapid:1 behavior:2 inspired:1 globally:1 freeman:4 actual:13 provided:1 notation:2 underlying:1 panel:2 moreover:1 medium:1 kaufman:1 interpreted:1 transformation:1 guarantee:2 sky:1 ti:8 exactly:2 prohibitively:1 grant:3 appear:1 positive:1 engineering:1 local:14 understood:1 modify:1 tends:1 consequence:5 frey:1 oxford:2 approximately:1 noteworthy:1 might:1 studied:1 mateo:1 equivalence:1 factorization:3 practical:6 unique:1 acknowledgment:1 implement:1 daad19:1 area:2 significantly:1 algorithmindependent:1 storage:1 context:1 optimize:1 restriction:2 equivalent:3 map:3 straightforward:1 regardless:1 simplicity:1 subgraphs:1 insight:5 rule:1 importantly:2 reparameterization:25 stability:1 updated:1 transmit:1 exact:20 us:1 element:2 roy:1 muri:1 cooperative:1 electrical:1 cycle:13 mjwain:1 ordering:1 removed:3 rescaled:1 valuable:1 substantial:1 convexity:1 complexity:1 comm:2 dynamic:1 ultimately:1 motivate:1 tight:2 purely:1 yuille:1 joint:1 various:3 distinct:2 describe:1 artificial:1 emerged:1 widely:1 larger:1 tightness:2 richardson:2 superscript:1 sequence:2 advantage:1 rr:1 product:5 neighboring:2 tu:1 loop:4 subgraph:1 mixing:1 intuitive:1 convergence:6 cluster:1 p:5 assessing:1 double:1 produce:1 converges:1 leave:1 kikuchi:6 derive:4 develop:2 depending:1 ac:1 lauritzen:1 strong:4 implies:3 tommi:2 closely:1 stochastic:2 enable:1 jst:4 implementing:1 suffices:1 elementary:1 extension:4 mjw:1 hold:2 lying:1 omitted:1 estimation:1 applicable:1 expose:1 correctness:2 minimization:1 reparameterizing:1 mit:8 gaussian:4 powering:1 rather:1 sel:2 factorizes:1 jaakkola:2 focus:2 notational:1 tech:1 contrast:1 inference:12 unlikely:1 compatibility:2 overall:1 among:1 development:1 constrained:5 initialize:1 marginal:11 field:1 identical:1 t2:3 report:1 quantitatively:1 intelligent:1 divergence:1 individual:2 geometry:1 consisting:1 attempt:1 freedom:1 interest:1 message:6 highly:1 severe:1 analyzed:1 edge:8 necessary:1 tree:46 desired:1 plotted:1 theoretical:3 assignment:1 loopy:2 cost:3 vertex:1 deduced:1 density:1 fundamental:1 probabilistic:1 decoding:3 iy:1 clifford:1 thesis:2 choose:3 positivity:1 potential:6 singleton:3 coding:1 sec:1 includes:2 rusmevichientong:1 modl:1 satisfy:2 jc:3 explicitly:2 performed:3 view:3 lab:1 analyze:2 doing:1 bayes:1 parallel:1 ttt:1 substructure:2 contribution:2 minimize:1 accuracy:1 efficiently:1 yield:2 correspond:2 t3:2 ssg:1 weak:2 bayesian:1 researcher:2 energy:9 minka:1 obvious:1 naturally:3 proof:1 pst:3 massachusetts:1 schedule:1 sophisticated:1 actually:1 reflected:2 wei:5 though:4 anderson:1 just:1 lastly:1 d:1 ei:1 propagation:9 continuity:1 reveal:1 effect:1 iteratively:1 leibler:1 laboratory:1 attractive:2 coincides:1 generalized:1 tt:1 demonstrate:1 tn:9 performs:1 reasoning:1 variational:2 recently:1 superior:1 functional:1 rl:1 physical:1 rnit:1 extend:1 interpretation:3 belong:2 marginals:25 refer:1 measurement:3 cambridge:1 ai:1 approx:1 consistency:5 grid:5 funded:1 stable:1 j:5 recent:1 showed:1 perspective:7 compound:1 certain:1 n00014:1 binary:1 onr:1 morgan:1 minimum:6 converge:3 july:1 ii:7 full:10 desirable:1 alan:1 faster:1 lin:1 y:1 basic:3 expectation:2 iteration:3 normalization:3 represent:1 fellowship:1 pseudomarginal:1 subject:1 tend:2 undirected:1 ee:1 revealed:1 gave:1 topology:1 idea:3 computable:1 whether:2 expression:3 reformulated:3 passing:2 generally:2 useful:2 involve:1 http:1 exist:1 arising:1 discrete:1 shall:2 express:1 group:2 tst:9 graph:37 sum:1 parameterized:1 uncertainty:1 place:2 throughout:3 family:1 looser:1 decision:1 bound:24 yielded:1 turbo:3 constraint:3 bp:42 ri:4 argument:2 reparameterize:1 optimality:1 martin:1 relatively:2 department:1 structured:5 according:2 poor:3 remain:1 lid:1 restricted:1 invariant:4 equation:4 remains:2 turn:1 loose:2 tractable:1 junction:2 operation:1 available:1 yedidia:2 apply:3 alternative:4 existence:2 original:4 denotes:5 graphical:6 xc:3 establish:1 february:2 unchanged:1 quantity:5 dependence:1 graphtheoretic:1 decoder:1 spanning:13 reason:1 willsky:2 induction:1 itu:2 code:2 index:2 illustration:2 minimizing:1 unfortunately:1 statement:1 info:2 rise:1 suppress:1 implementation:1 motivates:1 perform:3 allowing:1 upper:3 teh:1 observation:1 markov:1 t:11 january:2 reparameterized:1 extended:1 perturbation:1 arbitrary:9 lb:1 pair:2 subvector:1 specified:1 namely:3 tap:1 merges:1 established:1 pearl:1 nip:3 trans:2 address:1 appeared:1 hammersley:1 including:6 max:3 belief:8 wainwright:3 natural:3 difficulty:2 indicator:1 residual:1 recursion:1 technology:1 sev:2 prior:1 literature:1 review:1 geometric:1 afosr:1 embedded:8 mixed:2 acyclic:5 proven:1 remarkable:1 versus:1 consistent:3 pi:6 placed:1 supported:1 free:12 side:2 allow:1 institute:1 neighbor:1 emerge:1 van:1 f49620:1 computes:1 t5:1 san:1 welling:1 approximate:11 kullback:1 clique:5 global:3 conceptual:5 un:1 iterative:2 decomposes:1 bethe:9 nature:1 kschischang:1 expansion:1 excellent:3 complex:1 necessarily:1 arise:1 tl:4 en:2 lc:1 fails:1 explicit:1 lie:1 theorem:1 removing:1 xt:3 jt:3 symbol:1 x:7 intrinsic:3 exists:1 intractable:1 phd:2 notwithstanding:1 illustrates:2 t4:1 simply:1 nserc:1 partially:1 joined:1 trp:45 applies:1 corresponds:1 satisfies:1 ma:1 conditional:1 viewed:2 identity:1 change:1 determined:1 invariance:10 la:1 arises:1 pythagorean:1 phenomenon:1 ex:4 |
1,216 | 2,108 | Semi-Supervised MarginBoost
F. d'Alche-Buc
LIP6,UMR CNRS 7606,
Universite P. et M. Curie
75252 Paris Cedex, France
Yves Grandvalet
Heudiasyc, UMR CNRS 6599,
Universite de Technologie de Compiegne,
BP 20.529, 60205 Compiegne cedex, France
florence. [email protected]
Yves. [email protected]
Christophe Ambroise
Heudiasyc, UMR CNRS 6599,
Universite de Technologie de Compiegne,
BP 20.529, 60205 Compiegne cedex, France
Christophe A [email protected]
Abstract
In many discrimination problems a large amount of data is available but
only a few of them are labeled. This provides a strong motivation to
improve or develop methods for semi-supervised learning. In this paper,
boosting is generalized to this task within the optimization framework of
MarginBoost . We extend the margin definition to unlabeled data and
develop the gradient descent algorithm that corresponds to the resulting
margin cost function. This meta-learning scheme can be applied to any
base classifier able to benefit from unlabeled data. We propose here to
apply it to mixture models trained with an Expectation-Maximization
algorithm. Promising results are presented on benchmarks with different
rates of labeled data.
1
Introduction
In semi-supervised classification tasks, a concept is to be learnt using both labeled
and unlabeled examples. Such problems arise frequently in data-mining where the
cost of the labeling process can be prohibitive because it requires human help as in
video-indexing, text-categorization [12] and medical diagnosis. While some works
proposed different methods [16] to learn mixture models [12], [1], SVM [3], cotrained machines [5] to solve this task, no extension has been developed so far for
ensemble methods such as boosting [7, 6]. Boosting consists in building sequentially a linear combination of base classifiers that focus on the difficult examples.
For AdaBoost and extensions such as MarginBoost [10], this stage-wise procedure
corresponds to a gradient descent of a cost functional based on a decreasing function
of the margin, in the space of linear combinations of base classifiers.
We propose to generalize boosting to semi-supervised learning within the framework of optimization. We extend the margin notion to unlabeled data, derive the
corresponding criterion to be maximized, and propose the resulting algorithm called
Semi-Supervised MarginBoost (SSMBoost). This new method enhances our previ-
ous work [9] based on a direct plug-in extension of AdaBoost in the sense that all
the ingredients of the gradient algorithm such as the gradient direction and the
stopping rule are defined from the expression of the new cost function. Moreover,
while the algorithm has been tested using the mixtures of models [1], 55MBoost
is designed to combine any base classifiers that deals with both labeled and unlabeled data. The paper begins with a brief presentation of MarginBoost (section 2).
Then, in section 3, the 55MBoost algorithm is presented. Experimental results are
discussed in section 5 and we conclude in section 6.
2
Boosting with MarginBoost
Boosting [7, 6, 15] aims at improving the performance of any weak "base classifier" by linear combination. We focus here on normalized ensemble classifiers
gt E LinCH) whose normalized 1 coefficients are noted aT = I ~: I and each base
classifier with outputs in [-1, 1] is hT E 1{:
t
gt(x) =
L aThT(x)
(1)
T=l
Different contributions [13, 14],[8], [10] have described boosting within an optimization scheme, considering that it carries out a gradient descent in the space of linear
combinations of base functions. We have chosen the MarginBoost algorithm, a variant of a more general algorithm called Any Boost [10], that generalizes AdaBoost
and formally justifies the interpretation in terms of margin. If S is the training
sample {(Xi,Yi) , i = l..l}, MarginBoost, described in Fig. 1, minimizes the cost
functional C defined for any scalar decreasing function c of the margin p :
I
C(gt) =
L c(p(gt(Xi), Yi)))
(2)
i=l
Instead of taking exactly ht+l = - \1C(gt) which does not ensure that the resulting
function gt+! belongs to Lin(1{), ht+! is chosen such as the inner product 2 - <
\1C(gt), ht+l > is maximal. The equivalent weighted cost function to be maximized
can thus be expressed as :
JF = L Wt(i)Yiht+! (Xi)
(3)
iES
3
Generalizing MarginBoost to semi-supervised
classification
3.1
Margin Extension
For labeled data, the margin measures the quality of the classifier output. When no
label is observed, the usual margin cannot be calculated and has to be estimated.
A first estimation could be derived from the expected margin EypL(gt(X) , y). We
can use the output of the classifier (gt(x) + 1)/2 as an estimate of the posterior
probability P(Y = +llx). This leads to the following margin pi; which depends on
the input and is linked with the response of the classifier:
lOr>
0 and L1 norm is used for normalization:
2< f, 9 >=
LiE S
f(X;)g(Xi)
IOrl
= L~=l
Or
Let wo(i) = l/l , i = 1, ... ,l.
Let go(x) = 0
For t = 1 ... T (do the gradient descent):
1. Learn a gradient direction htH E 1i with a high value of
J{ = L,iEswt(i)YihtH(Xi)
2. Apply the stopping rule: if J{ ::::: L,iES Wt(i)Yigt(Xi) then return
gt else go on.
3. Choose a step-length for the obtained direction by a line-search or
by fixing it as a constant f
a ttlh t t')
4 . Add the new direction to obtain 9HI = (l a t I9t+
lattl l
5. Fix the weight distribution: Wt
+1
=
c'(p(9ttl(Xi),Yi))
2: jE S c'(p(9ttl(Xj),Yj))
Figure 1: MarginBoost algorithm (with L1 normalization of the combination coefficients)
Another way of defining the extended margin is to use directly the maximum a
posteriori estimate of the true margin. This MAP estimate depends on the sign of
the classifier output and provides the following margin definition pC; :
(5)
3.2
Semi-Supervised MarginBoost : generalization of marginBoost to
deal with unlabeled data
The generalization of the margin can be used to define an appropriate cost functional
for the semi-supervised learning task. Considering that the training sample S is now
divided into two disjoint subsets L for labeled data and U for unlabeled data, the
cost falls into two parts involving PL = P and PU:
(6)
iEL
iEU
The maximization of - < \lC(gt), htH > is equivalent to optimize the new quantity
JtS that falls now into two terms J{ = Jf + J? The first term one can be directly
obtained from equation (3) :
Jf = LWt(i).YihtH(Xi)
(7)
iEL
The second term, J? , can be expressed as following:
(8)
with the weight distribution
Wt
now defined as :
c'(pL(9t( Xi),Yi))
Wt
(z.) -_
{
if i E L
..
with
If z E U
IWt l
c'(PU(9t(Xi)))
IWt l
This expression of
product:
IWt
I=
2:=
Wt
(i)
(9)
iES
JP comes directly from differential calculus and the chosen inner
( )()
'VC gt Xi
if x = Xi and i E L
if x = x, and i E U
YiC'(Pd9t(Xi),Yi))
= { c'(p U (g t (x.))) apU(9t(Xi))
a9t( Xi)
t
(10)
0
Pu
Implementation of 55MBoost with margins pI[; and
requires their derivatives.
Let us notice that the "signed margin", pus, is not derivable at point O. However,
according to the results of convex analysis (see for instance [2]), it is possible to
define the "sub derivative' of Pus since it is a continuous and convex function. The
value of the sub derivative corresponds here to the average value of the right and
left derivatives.
apUS(gt(Xi)) = {sign(g(Xi))
agt (Xi)
0
if X :f": 0
if x = 0
(11)
And, for the "squared margin" Pu 9 , we have:
apu 9 (gt(Xi)) = 2g(Xi)
agt(Xi)
(12)
This completes the set of ingredients that must be incorporated into the algorithm
of Fig. 1 to obtain 55MBoost.
4
Base Classifier
The base classifier should be able to make use of the unlabeled data provided by the
boosting algorithm. Mixture models are well suited for this purpose, as shown by
their extensive use in clustering. Hierarchical mixtures provide flexible discrimination tools, where each conditional distribution f(xlY = k) is modelled by a mixture
of components [4]. At the high level, the distribution is described by
K
f(x; if? =
2:= Pk!k (x; Ok)
,
(13)
k=l
where K is the number of classes, Pk are the mixing proportions, Ok the conditional
distribution parameters, and if> denotes all parameters {Pk; 0df=l. The high-level
description can also be expressed as a low-level mixture of components, as shown
here for binary classification:
Kl
f(x;if? =
K2
2:= PkJkl(X;Okl) + 2:= Pk2!k2(X;Ok2)
(14)
With this setting, the EM algorithm is used to maximize the log-likelihood with
respect to if> considering the incomplete data is {Xi, Yi}~= l and the missing data
is the component label Cik, k = 1, ... , K 1 + K2 [11]. An original implementation
of EM based on the concept of possible labels [1] is considered here. It is well
adapted to hierarchical mixtures, where the class label Y provides a subset of possible
components. When Y = 1 the first Kl modes are possible, when Y = -1 the last
K2 modes are possible, and when an example is unlabeled, all modes are possible.
A binary vector Zi E {0,1}(Kl+ K2) indicates the components from which feature
vector Xi may have been generated, in agreement with the assumed mixture model
and the (absence of) label Yi. Assuming that the training sample {Xi, Zi }i=l is i.i.d ,
the weighted log-likelihood is given by
I
L(<I> ;{Xi,zdi=l
= LWt(i) log (j(Xi,zi;<I?) ,
(15)
i=l
where Wt(i) are provided by boosting at step t. L is maximized using the following
EM algorithm:
E-Step Compute the expectation of L( <I>; {Xi , zdi=l) conditionally to {Xi , zdi=l
and the current value of <I> (denoted <I>q):
Kl+K2
L
Wt(i)Uik log (Pk!k(Xi; Ok))
i=l k=l
ZikPk!k(Xi; Ok)
L? ZUP?!?( Xi; O?)
I
L
with Uik
(16)
M-Step Maximize Q(<I>I<I>q) with respect to <I>.
Assuming that each mode k follows a normal distribution with mean ILk'
and covariance ~k ' <I>q+l = {ILk+! ; ~k+!;Pk+l}f~iK2 is given by:
(17)
(18)
5
Experimental results
Tests of the algorithm are performed on three benchmarks of the boosting literature:
twonorm and ringnorm [6] and banana [13]. Information about these datasets and
the results obtained in discrimination are available at www.first.gmd.de/-raetsch/
10 different samples were used for each experiment.
We first study the behavior of 55MBoost according the evolution of the test error
with increasing rates of unlabeled data (table 1). We consider five different settings
where 0%, 50%, 75%, 90% and 95% of labels are missing. 55MB is tested for the
margins P~ and Pu with c(x) = exp( -x). It is compared to mixture models and
AdaBoost. 55MBoost and AdaBoost are trained identically, the only difference
being that AdaBoost is not provided with missing labels.
Both algorithms are run for T = 100 boosting steps, without special care of overfitting. The base classifier (called here base(EM)) is a hierarchical mixture model with
an arbitrary choice of 4 modes per class but the algorithm (which may be stalled
in local minima) is restarted 100 times from different initial solutions, and the best
final solution (regarding training error rate) is selected. We report mean error rates
together with the lower and upper quartiles in table 1. For sake of space, we did
not display the results obtained without missing labels: in this case, AdaBoost and
55MBoost behave nearly identically and better than EM only for Banana.
For rates of unlabeled data inferior to 95% , 55MBoost beats slightly AdaBoost
for Ringnorm and Twonorm (except for 75%) but is not able to do as well as
Table 1: Mean error rates (in %) and interquartiles obtained with 4 different percentages
of unlabeled data for mixture models base(EM), AdaBoost and 55MBoost.
Ringnorm
50%
75%
90%
95%
base(EM)
AdaBoost
55MBoost pS
55MBoost pg
Twonorm
2.1 [ 1.7,
1.8[ 1.6,
1. 7[ 1.5,
1. 7[ 1.6,
50%
2.1]
2.0]
1.8]
1.8]
4.3[ 1.9,
3.1[ 1.9,
2.0 [ 1.5,
2.O[ 1.4,
75%
5.7]
4.1]
2.4]
2.5]
9.5 [ 2.7,12.0]
11.5[ 4.2 ,12.1]
3.7[ 2.1, 4.8]
4.5 [ 2.2, 3.6]
90%
23.7 [1 4.5,27.0]
28.7[11.5,37.6]
6.9[ 5.6,10.7]
8.1 [ 4.2, 9.0]
95%
base(EM)
AdaBoost
55MBoost pS
55MBoost pg
Banana
3.2 [ 2.7,
3.2[ 2.9,
2.7[ 2.5,
2.7[ 2.5,
50%
3.1]
3.2]
2.9]
2.8]
6.5[ 3.0,
3.2[ 3.0,
3.4 [ 2.8,
3.4 [ 2.8,
75%
9.0] 20.6[10.3,22.5] 24.8[18.3,31.9]
3.5] 11.0[ 5.2,14.2] 38.9[29.4,50.0]
4.3] 10.1 [ 5.8,13.6] 20.4[11.9,32.3]
4.2] 11.0[ 5.6,16.2] 21.1 [1 2.5,30.8]
90%
95%
base(EM)
18.2[16.7,18.6] 21.8[18.0,25.0] 26.1[20.7,29.8] 31.7[23.8,35.8]
AdaBoost
12.6[11.7,13.1] 15.2 [13.0,16.8] 22.1 [18.0,24.3] 37.5 [32.2,42.2]
55MBoost pS 13.3 [1 2.7,14.3] 17.0[15.3,17.8] 22.2[18.0,28.0] 28.3 [20.2,35.2]
55MBoost pg 13.3[12.8,14.2] 16.9[15.6,17.8] 22.8[18.3,29.3] 28.6 [21.5,34.2]
AdaBoost on Banana data. One possible explanation is that the discrimination
frontiers involved in the banana problem are so complex that the labels really bring
crucial informations and thus adding unlabeled data does not help in such a case.
Pu
obtains
Nevertheless, at rate 95% which is the most realistic situation, the margin
the minimal error rate for each of the three problems. It shows that it is worth
boosting and using unlabeled data.
As there is no great difference between the two proposed margins, we conducted
further experiments using only the
Pu'
Second, in order to study the relation between the presence of noise in the dataset
and the ability of 55MBoost to enhance generalization performance, we draw in
Fig. 2, the test errors obtained for problems with different values of Bayes error
when varying the rate of labeled examples. We see that even for difficult tasks (very
noisy problems), the degradation in performance for large subsets of unlabeled data
is still low. This reflects some consistency in the behavior of our algorithm.
Third, we test the sensibility of 55MBoost to overfitting. Overfitting can usually
be avoided by techniques such as early stopping, softenizing of the margin ([1 3],
[14]) or using an adequate margin function such as 1 - tanh(p) instead of exp( -p)
[10]. Here we keep using c = exp and ran 55MBoost with a maximal number of
step T = 1000 with 95% of unlabeled data. Of course, this does not correspond to
a realistic use of boosting in practice but it allows to check if the algorithm behaves
consistently in terms of gradient steps number. It is remarkable that no overfitting
is observed and in the Twonorm case (see Fig. 3), the test error still decreases
! We also observe that the standard error deviation is reduced at the end of the
process. For the banana problem (see Fig. 3 b.), we observe a stabilization near
the step t = 100. A massive presence of unlabeled data implies thus a regularizing
effect.
Bayes error:;;; 2.3%
Bayes error:;;; 15 .7%
Bayes error:;;; 3 1 .2%
50
40
20
10
?0L---~,7
0 --~20
7---~~
7----4~0--~5~0--~6=0--~7=0----8=0----9=0~~
'00
Rate of missing labels (%)
Figure 2: Consistency of the 55MBoost behavior: evolution of test error versus the missing
labels rate with respect to various Bayes error (twonorm ).
Mean (Error Test) +/- 1 std
70
70
Mean (Error test)
60'
,
I~
Mean of Error Test +/- std
Mean of Error test
I
60
,
~ 50
I
i \!), \
\
0: 40"
"' \
"
~_~~
~",~~,.~.
'-.
-.I~"
__
-r~/_
~
~~--
10
oL-~
o
__
~
~
__ _ __
~
~
-'-",---.- - --/----_ L_ __ L_ _~_ _~_ _~_ _L_~
~
~
~
~
Steptofgradient descent(boosting process}
~
~
_
?OL-~'OO~-2~OO~~3~OO--~400~~500~-=~~~7~OO~~8=OO--~~~~'~
Step t of gradient descent
Figure 3: Evolution of Test error with respect to maximal number T of iterations with
95% of missing labels (Two norm and Banana).
6
Conclusion
MarginBoost algorithm has been extended to deal with both labeled and unlabeled
data. Results obtained on three classical benchmarks of boosting litterature show
that it is worth using additional information conveyed by the patterns alone. No
overfitting was observed during processing 55MBoost on the benchmarks when
95% of the labels are missing: this should mean that the unlabeled data should
playa regularizing role in the ensemble classifier during the boosting process. After
applying this method to a large real dataset such as those of text-categorization,
our future works on this theme will concern the use of the extended margin cost
function on the base classifiers itself like multilayered perceptrons or decision trees.
Another approach could also be conducted from the more general framework of
AnyBoost that optimize any differential cost function.
References
[1] C. Ambroise and G. Govaert. EM algorithm for partially known labels. In IFCS 2000,
july 2000.
[2] J.-P. Aubin. L 'analyse non lineaire et ses applications d l'economie. Masson , 1984.
[3] K P. Bennett and A. Demiriz. Semi-supervised support vector machines. In D. Cohn,
M. Kearns, and S. Solla, editors, Advances in Neural Information Processing Systems,
pages 368-374. MIT Press, 1999.
[4] C.M. Bishop and M.E. Tipping. A hierarchical latent variable model for data vizualization. IEEE PAMI, 20:281- 293, 1998.
[5] A. Blum and Tom Mitchell. Combining labeled and unlabeled data with co-training.
In Proceedings of th e 1998 Conference on Computational Learning Th eory, July 1998.
[6] L. Breiman. Prediction games and arcing algorithms. Technical Report 504, Statistics
Department , University of California at Berkeley, 1997.
[7] Y . Freund and R. E. Schapire. Experiments with a new boosting algorithm. In
Machin e Learning: Proceedings of th e Thirteenth International Conference, pages
148- 156. Morgan Kauffman, 1996.
[8] J. Friedman, T. Hastie, and R. Tibshirani. Additive logistic regression: a statistical
view of boosting. The Annals of Statistics, 28(2):337- 407, 2000.
[9] Y. Grandvalet, F. d'Alche Buc, and C. Ambroise. Boosting mixture models for semisupervised learning. In ICANN 2001 , august 200l.
[10] L. Mason , J. Baxter, P. L. Bartlett, and M. Frean. Functional gradient techniques
for combining hypotheses. In Advances in Large Margin Classifiers. MIT, 2000.
[11] G.J. McLachlan and T. Krishnan. Th e EM algorithm and extensions. Wiley, 1997.
[12] K Nigam, A. K McCallum, S. Thrun, and T. Mitchell. Text classification from
labeled and unlabeled documents using EM. Machine learning, 39(2/3):135- 167,
2000.
[13] G. Riitsch, T. Onoda, and K-R. Muller. Soft margins for AdaBoost. Technical report,
Department of Computer Science, Royal Holloway, London , 1998.
[14] G. Riitsch, T. Onoda, and K-R. Muller. Soft margins for AdaBoost. Machine Learning, 42(3):287- 320, 200l.
[15] R. E. Schapire, Y. Freund, P. Bartlett, and W. S. Lee. Boosting the margin: A
new explanation for the effectiveness of voting methods. Th e Annals of Statistics,
26(5):1651- 1686, 1998.
[16] Matthias
Seeger.
Learning
with
data,www.citeseer.nj.nec.com/seegerOllearning.html.
labeled
and
unlabeled
| 2108 |@word norm:2 proportion:1 calculus:1 covariance:1 pg:3 citeseer:1 carry:1 initial:1 document:1 riitsch:2 current:1 com:1 must:1 additive:1 realistic:2 designed:1 discrimination:4 alone:1 prohibitive:1 selected:1 mccallum:1 provides:3 boosting:20 lor:1 five:1 direct:1 differential:2 anyboost:1 consists:1 combine:1 expected:1 behavior:3 frequently:1 ol:2 utc:2 decreasing:2 considering:3 increasing:1 begin:1 provided:3 moreover:1 ttl:2 minimizes:1 developed:1 sensibility:1 nj:1 berkeley:1 voting:1 exactly:1 classifier:17 k2:6 medical:1 local:1 dalche:1 pami:1 signed:1 umr:3 ringnorm:3 co:1 apu:2 yj:1 practice:1 procedure:1 cannot:1 unlabeled:22 applying:1 optimize:2 equivalent:2 map:1 www:2 missing:8 go:2 masson:1 convex:2 hds:2 zdi:3 rule:2 pk2:1 ambroise:3 notion:1 annals:2 machin:1 massive:1 hypothesis:1 agreement:1 std:2 labeled:11 observed:3 role:1 solla:1 decrease:1 ran:1 technologie:2 trained:2 alche:2 various:1 london:1 labeling:1 whose:1 solve:1 tested:2 s:1 ability:1 statistic:3 economie:1 analyse:1 noisy:1 itself:1 final:1 demiriz:1 matthias:1 propose:3 product:2 maximal:3 fr:3 mb:1 combining:2 mixing:1 description:1 p:3 categorization:2 help:2 derive:1 develop:2 oo:5 frean:1 fixing:1 strong:1 come:1 xly:1 implies:1 direction:4 quartile:1 vc:1 stabilization:1 human:1 fix:1 generalization:3 really:1 aubin:1 extension:5 pl:2 frontier:1 considered:1 normal:1 exp:3 great:1 early:1 purpose:1 estimation:1 label:14 tanh:1 lip6:2 tool:1 weighted:2 reflects:1 mclachlan:1 mit:2 aim:1 breiman:1 varying:1 arcing:1 heudiasyc:2 focus:2 derived:1 consistently:1 likelihood:2 indicates:1 check:1 seeger:1 sense:1 posteriori:1 stopping:3 cnrs:3 relation:1 france:3 classification:4 flexible:1 html:1 denoted:1 special:1 nearly:1 future:1 report:3 few:1 friedman:1 mining:1 mixture:13 pc:1 tree:1 incomplete:1 minimal:1 instance:1 soft:2 maximization:2 cost:10 deviation:1 subset:3 conducted:2 learnt:1 international:1 twonorm:5 lee:1 enhance:1 together:1 jts:1 squared:1 choose:1 derivative:4 return:1 de:5 coefficient:2 depends:2 performed:1 view:1 linked:1 bayes:5 ifc:1 florence:1 curie:1 contribution:1 yves:2 ensemble:3 maximized:3 correspond:1 generalize:1 weak:1 modelled:1 worth:2 definition:2 involved:1 universite:3 dataset:2 mitchell:2 cik:1 ok:4 tipping:1 supervised:9 adaboost:15 response:1 tom:1 stage:1 cohn:1 mode:5 logistic:1 quality:1 semisupervised:1 building:1 effect:1 concept:2 normalized:2 true:1 evolution:3 deal:3 conditionally:1 during:2 game:1 inferior:1 noted:1 criterion:1 generalized:1 compiegne:4 l1:2 bring:1 wise:1 behaves:1 functional:4 jp:1 extend:2 discussed:1 interpretation:1 raetsch:1 llx:1 consistency:2 gt:14 base:16 add:1 playa:1 pu:9 posterior:1 belongs:1 meta:1 binary:2 christophe:2 ous:1 yi:7 muller:2 morgan:1 minimum:1 additional:1 care:1 maximize:2 july:2 semi:9 technical:2 plug:1 lin:1 divided:1 y:3 prediction:1 variant:1 involving:1 regression:1 expectation:2 df:1 iteration:1 normalization:2 thirteenth:1 else:1 completes:1 crucial:1 cedex:3 effectiveness:1 near:1 presence:2 identically:2 baxter:1 krishnan:1 xj:1 zi:3 hastie:1 inner:2 regarding:1 expression:2 bartlett:2 wo:1 adequate:1 amount:1 gmd:1 reduced:1 eory:1 schapire:2 percentage:1 notice:1 sign:2 estimated:1 disjoint:1 tibshirani:1 per:1 stalled:1 diagnosis:1 nevertheless:1 blum:1 ht:4 run:1 draw:1 decision:1 hi:1 iwt:3 display:1 ilk:2 adapted:1 bp:2 sake:1 department:2 according:2 combination:5 slightly:1 em:12 indexing:1 equation:1 end:1 available:2 generalizes:1 apply:2 observe:2 hierarchical:4 appropriate:1 original:1 denotes:1 clustering:1 ensure:1 classical:1 quantity:1 usual:1 govaert:1 enhances:1 gradient:10 thrun:1 assuming:2 length:1 difficult:2 implementation:2 litterature:1 upper:1 marginboost:13 datasets:1 benchmark:4 descent:6 behave:1 beat:1 defining:1 extended:3 incorporated:1 situation:1 banana:7 arbitrary:1 august:1 paris:1 kl:4 extensive:1 california:1 boost:1 able:3 usually:1 pattern:1 kauffman:1 royal:1 video:1 explanation:2 scheme:2 improve:1 brief:1 text:3 literature:1 freund:2 okl:1 versus:1 ingredient:2 remarkable:1 conveyed:1 editor:1 previ:1 grandvalet:3 pi:2 course:1 ok2:1 last:1 l_:3 ik2:1 fall:2 taking:1 benefit:1 calculated:1 avoided:1 far:1 obtains:1 derivable:1 buc:2 keep:1 sequentially:1 overfitting:5 conclude:1 assumed:1 xi:31 search:1 continuous:1 latent:1 yic:1 table:3 promising:1 learn:2 onoda:2 nigam:1 improving:1 hth:2 complex:1 did:1 pk:5 icann:1 multilayered:1 motivation:1 noise:1 arise:1 atht:1 fig:5 je:1 uik:2 wiley:1 lc:1 sub:2 theme:1 lie:1 third:1 yiht:1 bishop:1 mason:1 svm:1 concern:1 adding:1 agt:2 iel:2 nec:1 justifies:1 margin:28 lwt:2 suited:1 generalizing:1 expressed:3 partially:1 scalar:1 restarted:1 corresponds:3 conditional:2 presentation:1 jf:3 absence:1 bennett:1 except:1 wt:8 degradation:1 kearns:1 called:3 experimental:2 perceptrons:1 formally:1 holloway:1 support:1 regularizing:2 |
1,217 | 2,109 | Categorization by Learning
and Combining Object Parts
Bernd Heisele Thomas Serre Massimiliano Pontil Thomas Vetter Tomaso Poggio
Center for Biological
and Computational Learning, M.I.T., Cambridge, MA, USA
Honda R&D Americas, Inc., Boston, MA, USA
Department of Information Engineering, University of Siena, Siena, Italy
Computer Graphics Research Group, University of Freiburg, Freiburg, Germany
heisele,serre,tp @ai.mit.edu [email protected] [email protected]
Abstract
We describe an algorithm for automatically learning discriminative components of objects with SVM classifiers. It is based on growing image
parts by minimizing theoretical bounds on the error probability of an
SVM. Component-based face classifiers are then combined in a second
stage to yield a hierarchical SVM classifier. Experimental results in face
classification show considerable robustness against rotations in depth and
suggest performance at significantly better level than other face detection
systems. Novel aspects of our approach are: a) an algorithm to learn
component-based classification experts and their combination, b) the use
of 3-D morphable models for training, and c) a maximum operation on
the output of each component classifier which may be relevant for biological models of visual recognition.
1 Introduction
We study the problem of automatically synthesizing hierarchical classifiers by learning discriminative object parts in images. Our motivation is that most object classes (e.g. faces,
cars) seem to be naturally described by a few characteristic parts or components and their
geometrical relation. Greater invariance to viewpoint changes and robustness against partial occlusions are the two main potential advantages of component-based approaches compared to a global approach.
The first challenge in developing component-based systems is how to choose automatically
a set of discriminative object components. Instead of manually selecting the components,
it is desirable to learn the components from a set of examples based on their discriminative
power and their robustness against pose and illumination changes. The second challenge is
to combine the component-based experts to perform the final classification.
2 Background
Global approaches in which the whole pattern of an object is used as input to a single
classifier were successfully applied to tasks where the pose of the object was fixed. In [6]
Haar wavelet features are used to detect frontal and back views of pedestrians with an SVM
classifier. Learning-based systems for detecting frontal faces based on a gray value features
are described in [14, 13, 10, 2].
Component-based techniques promise to provide more invariance since the individual components vary less under pose changes than the whole object. Variations induced by pose
changes occur mainly in the locations of the components. A component-based method for
detecting faces based on the empirical probabilities of overlapping rectangular image parts
is proposed in [11]. Another probabilistic approach which detects small parts of faces is
proposed in [4]. It uses local feature extractors to detect the eyes, the corner of the mouth,
and the tip of the nose. The geometrical configuration of these features is matched with
a model configuration by conditional search. A related method using statistical models is
published in [9]. Local features are extracted by applying multi-scale and multi-orientation
filters to the input image. The responses of the filters on the training set are modeled as
Gaussian distributions. In [5] pedestrian detection is performed by a set of SVM classifiers
each of which was trained to detect a specific part of the human body.
In this paper we present a technique for learning relevant object components. The technique
starts with a set of small seed regions which are gradually grown by minimizing a bound
on the expected error probability of an SVM. Once the components have been determined,
we train a system consisting of a two-level hierarchy of SVM classifiers. First, component
classifiers independently detect facial components. Second, a combination classifier learns
the geometrical relation between the components and performs the final detection of the
object.
3 Learning Components with Support Vector Machines
3.1 Linear Support Vector Machines
Linear SVMs [15] perform pattern recognition for two-class problems by determining the
separating hyperplane with maximum distance to the closest points in the training set.
These points are called support vectors. The decision function of the SVM has the form:
!
#"%$ $
(1)
where is the number of data points and
is the class label of the data point
& . The coefficients are the solution of a quadratic programming
problem. The margin
'
is the distance of the support vectors to the hyperplane, it is given by:
'
$
( ) *
(2)
The margin is an indicator of the separability of the data. In fact, the expected error probability of the SVM, +-,.0/1/ , satisfies the following bound [15]:
+2,3.4/5/%6
where
9
$
+87:' 9<;
;>=
(3)
is the diameter of the smallest sphere containing all data points in the training set.
3.2 Learning Components
Our method automatically determines rectangular components from a set of object images.
The algorithm starts with a small rectangular component located around a pre-selected
point in the object image (e.g. for faces this could be the center of the left eye). The component is extracted from each object image to build a training set of positive examples. We
also generate a training set of background patterns that have the same rectangular shape as
the component. After training an SVM on the component data we estimate the performance
of the SVM based on the upper bound on the error probability. According to Eq. (3) we
calculate:
' 9<;
;
*
(4)
As shown in [15] this quantity can be computed by solving a quadratic programming problem. After determining we enlarge the component by expanding the rectangle by one
pixel into one of the four directions (up, down, left, right). Again, we generate training
data, train an SVM and determine . We do this for expansions into all four directions
and finally keep the expansion which decreases the most. This process is continued until
the expansions into all four directions lead to an increase of . In order to learn a set of
components this process can be applied to different seed regions.
4 Learning Facial Components
Extracting face patterns is usually a tedious and time-consuming work that has to be done
manually. Taking the component-based approach we would have to manually extract each
single component from all images in the training set. This procedure would only be feasible
for a small number of components. For this reason we used textured 3-D head models [16]
to generate the training data. By rendering the 3-D head models we could automatically
generate large numbers of faces in arbitrary poses and with arbitrary illumination. In addition to the 3-D information we also knew the 3-D correspondences for a set of reference
points shown in Fig. 1a). These correspondences allowed us to automatically extract facial
components located around the reference points. Originally we had seven textured head
models acquired by a 3-D scanner. Additional head models were generated by 3-D morph"
ing between
all pairs of the original head models. The heads were rotated between
and
in depth. The faces were illuminated by ambient light and a single directional
" light
and
pointing
towards
the
center
of
the
face.
The
position
of
the
light
varied
between
in azimuth and between
and
in elevation. Overall, we generated 2,457 face
images of size 58 58. Some examples of synthetic face images used for training are shown
in Fig. 1b).
The negative training set initially consisted of 10,209 58 58 non-face patterns randomly
extracted from 502 non-face images. We then applied bootstrapping to enlarge the training
data by non-face patterns that look similar to faces. To do so we trained a single linear
SVM classifier and applied it to the previously used set of 502 non-face images. The false
positives (FPs) were added to the non-face training data to build the final training set of
size 13,654.
We started with fourteen manually selected seed regions of size 5 5. The resulting components were located around the eyes (17 17 pixels), the nose (15 20 pixels), the mouth
$
(31 15 pixels), the cheeks (21 20 pixels), the lip (13 16 pixels), the nostrils (
pixels), the corners of the mouth (18 11 pixels), the eyebrows (19 15 pixels), and the
bridge of the nose (18 16 pixels).
a)
b)
Figure 1: a) Reference points on the head models which were used for 3-D morphing and
automatic extraction of facial components. b) Examples of synthetic faces.
5 Combining Components
An overview of our two-level component-based classifier is shown in Fig. 2. On the first
level the component classifiers independently detect components of the face. Each classifier was trained on a set of facial components and on a set of non-face patterns generated
from the training set described in Section 4. On the second level the combination classifier performs the detection of the face based on the outputs of the component classifiers.
The maximum real-valued outputs of each component classifier within rectangular search
regions around the expected positions of the components are used as inputs to the combination classifier. The size of the search regions was estimated from the mean and the standard
deviation of the locations of the components in the training images. The maximum operation is performed both during training and at run-time. Interestingly it turns out to be
similar to the key pooling mechanism postulated in a recent model of object recognition in
the visual cortex [8]. We also provide the combination classifier with the precise positions
of the detected components relative to the upper left corner of the 58
58 window. Overall
we have three values per component classifier that are propagated to the combination classifier: the maximum output of the component classifier and the & - image coordinates of
the maximum.
Left Eye
Eye
Left
expert:
expert:
Linear SVM
SVM
Linear
*
*Outputs of component
experts: bright intensities
indicate high confidence.
(O1 , X 1 , Y1 )
.
.
.
Nose expert:
expert:
Nose
Linear SVM
SVM
Linear
(O1 , X 1 , Y1 ,..., O14 , X 14 , Y14 )
*
(Ok , X k , Yk )
.
.
.
Mouth
Mouth
expert:
expert:
Linear SVM
SVM
Linear
1. Shift 58x58 window
over input image
2. Shift component
experts over
58x58 window
Combination
Combination
classifier:
classifier:
Linear SVM
SVM
Linear
*
(O14 , X 14 , Y14 )
3. For each component k,
determine its maximum
output within a search
region and its location:
(Ok , X k , Yk )
4. Final decision:
face / background
Figure 2: System overview of the component-based classifier.
6 Experiments
In our experiments we compared the component-based system to global classifiers. The
component system consisted of fourteen linear SVM classifiers for detecting the components and a single linear SVM as combination classifier. The global classifiers were a single
linear SVM and a single second-degree polynomial SVM both trained on the gray values
of the whole face pattern. The training data for these three classifiers consisted of 2,457
synthetic gray face images and 13,654 non-face gray images
" of size
58 58. The positive
test set consisted of 1,834 faces rotated between about
and
in depth. The faces
were manually extracted from the CMU PIE database [12]. The negative test set consisted
of 24,464 difficult non-face patterns that were collected by a fast face detector [3] from
web images. The FP rate was calculated relative to the number of non-face test images.
Because of the resolution required by the component-based system, a direct comparison
with other published systems on the standard MIT-CMU test set [10] was impossible. For
an indirect comparison, we used a second-degree polynomial SVM [2] which was trained
on a large set of 19 19 real face images. This classifier performed amongst the best face
detection systems on the MIT-CMU test set. The ROC curves in Fig. 3 show that the
component-based classifier is significantly better than the three global classifiers. Some
detection results generated by the component system are shown in Fig. 4.
Figure 3: Comparison between global classifiers and the component-based classifier.
Figure 4: Faces detected by the component-based classifier.
A natural question that arises is about the role of geometrical information. To answer this
question?which has relevant implications for models of cortex?we tested another system in
which the combination classifier receives as inputs only the output of each component classifier but not the position of its maximum. As shown in Fig. 5 this system still outperforms
the whole face systems but it is worse than the system with position information.
Figure 5: Comparison between a component-based classifier trained with position information and a component-based classifier without position information.
7 Open Questions
An extension under way of the component-based approach to face identification is already
showing good performances [1]. Another natural generalization of the work described here
involves the application of our system to various classes of objects such as cars, animals,
and people. Still another extension regards the question of view-invariant object detection. As suggested by [7] in a biological context and demonstrated recently by [11] in
machine vision, full pose invariance in recognition tasks can be achieved by combining
view-dependent classifiers. It is interesting to ask whether the approach described here
could also be used to learn which views are most discriminative and how to combine them
optimally. Finally, the role of geometry and in particular how to compute and represent
position information in biologically plausible networks, is an important open question at
the interface between machine and biological vision.
References
[1] B. Heisele, P. Ho, and T. Poggio. Face recognition with support vector machines:
global versus component-based approach. In Proc. 8th International Conference on
Computer Vision, Vancouver, 2001.
[2] B. Heisele, T. Poggio, and M. Pontil. Face detection in still gray images. A.I.
memo 1687, Center for Biological and Computational Learning, MIT, Cambridge,
MA, 2000.
[3] B. Heisele, T. Serre, S. Mukherjee, and T. Poggio. Feature reduction and hierarchy
of classifiers for fast object detection in video images. In Proc. IEEE Conference on
Computer Vision and Pattern Recognition, Hawaii, 2001.
[4] T. K. Leung, M. C. Burl, and P. Perona. Finding faces in cluttered scenes using
random labeled graph matching. In Proc. International Conference on Computer
Vision, pages 637?644, Cambridge, MA, 1995.
[5] A. Mohan, C. Papageorgiou, and T. Poggio. Example-based object detection in images by components. In IEEE Transactions on Pattern Analysis and Machine Intelligence, volume 23, pages 349?361, April 2001.
[6] C. Papageorgiou and T. Poggio. A trainable system for object detection. In International Journal of Computer Vision, volume 38, 1, pages 15?33, 2000.
[7] T. Poggio and S. Edelman. A network that learns to recognize 3-D objects. Nature,
343:163?266, 1990.
[8] M. Riesenhuber and T. Poggio. Hierarchical models of object recognition in cortex.
Nature Neuroscience, 2(11):1019?1025, 1999.
[9] T. D. Rikert, M. J. Jones, and P. Viola. A cluster-based statistical model for object
detection. In Proc. IEEE Conference on Computer Vision and Pattern Recognition,
volume 2, pages 1046?1053, Fort Collins, 1999.
[10] H. A. Rowley, S. Baluja, and T. Kanade. Rotation invariant neural network-based face
detection. Computer Science Technical Report CMU-CS-97-201, CMU, Pittsburgh,
1997.
[11] H. Schneiderman and T. Kanade. A statistical method for 3d object detection applied to faces and cars. In Proc. IEEE Conference on Computer Vision and Pattern
Recognition, pages 746?751, 2000.
[12] T. Sim, S. Baker, and M. Bsat. The CMU pose, illumination, and expression (PIE)
database of human faces. Computer Science Technical Report 01-02, CMU, 2001.
[13] K.-K. Sung. Learning and Example Selection for Object and Pattern Recognition.
PhD thesis, MIT, Artificial Intelligence Laboratory and Center for Biological and
Computational Learning, Cambridge, MA, 1996.
[14] R. Vaillant, C. Monrocq, and Y. Le Cun. An original approach for the localisation of
objects in images. In International Conference on Artificial Neural Networks, pages
26?30, 1993.
[15] V. Vapnik. Statistical learning theory. John Wiley and Sons, New York, 1998.
[16] T. Vetter. Synthesis of novel views from a single face. International Journal of
Computer Vision, 28(2):103?116, 1998.
| 2109 |@word polynomial:2 tedious:1 open:2 reduction:1 configuration:2 selecting:1 interestingly:1 outperforms:1 john:1 shape:1 intelligence:2 selected:2 detecting:3 location:3 honda:1 direct:1 fps:1 edelman:1 combine:2 acquired:1 expected:3 tomaso:1 growing:1 multi:2 detects:1 automatically:6 window:3 matched:1 baker:1 finding:1 vaillant:1 bootstrapping:1 sung:1 classifier:43 positive:3 engineering:1 local:2 procedure:1 heisele:5 pontil:3 empirical:1 significantly:2 matching:1 pre:1 vetter:3 confidence:1 suggest:1 selection:1 context:1 applying:1 impossible:1 demonstrated:1 center:5 independently:2 cluttered:1 rectangular:5 resolution:1 continued:1 variation:1 coordinate:1 hierarchy:2 programming:2 us:1 recognition:10 located:3 mukherjee:1 database:2 labeled:1 role:2 calculate:1 region:6 decrease:1 yk:2 rowley:1 trained:6 solving:1 textured:2 indirect:1 various:1 america:1 grown:1 train:2 massimiliano:1 fast:2 describe:1 detected:2 artificial:2 valued:1 plausible:1 tested:1 final:4 advantage:1 relevant:3 combining:3 cluster:1 categorization:1 rotated:2 object:25 pose:7 sim:1 eq:1 c:1 involves:1 indicate:1 direction:3 filter:2 human:2 generalization:1 elevation:1 biological:6 extension:2 scanner:1 around:4 seed:3 pointing:1 vary:1 smallest:1 proc:5 label:1 bridge:1 successfully:1 mit:5 gaussian:1 mainly:1 detect:5 dependent:1 leung:1 initially:1 y14:2 relation:2 perona:1 germany:1 pixel:10 overall:2 classification:3 orientation:1 animal:1 once:1 extraction:1 enlarge:2 manually:5 look:1 jones:1 report:2 few:1 randomly:1 recognize:1 individual:1 geometry:1 occlusion:1 consisting:1 detection:14 localisation:1 light:3 implication:1 ambient:1 partial:1 poggio:8 facial:5 theoretical:1 tp:1 deviation:1 azimuth:1 graphic:1 optimally:1 morph:1 answer:1 synthetic:3 combined:1 international:5 probabilistic:1 tip:1 synthesis:1 again:1 thesis:1 containing:1 choose:1 worse:1 corner:3 hawaii:1 expert:10 potential:1 de:1 coefficient:1 inc:1 pedestrian:2 postulated:1 performed:3 view:5 start:2 bright:1 characteristic:1 yield:1 directional:1 identification:1 informatik:1 published:2 detector:1 against:3 nostril:1 naturally:1 propagated:1 ask:1 car:3 back:1 ok:2 originally:1 response:1 april:1 done:1 stage:1 until:1 receives:1 web:1 overlapping:1 gray:5 usa:2 serre:3 consisted:5 burl:1 laboratory:1 during:1 freiburg:3 performs:2 interface:1 geometrical:4 image:24 novel:2 recently:1 rotation:2 overview:2 fourteen:2 volume:3 cambridge:4 ai:1 automatic:1 had:1 cortex:3 morphable:1 closest:1 recent:1 italy:1 greater:1 additional:1 determine:2 full:1 desirable:1 ing:2 technical:2 sphere:1 vision:9 cmu:7 represent:1 achieved:1 background:3 addition:1 induced:1 pooling:1 unisi:1 seem:1 extracting:1 rendering:1 shift:2 whether:1 expression:1 york:1 svms:1 diameter:1 generate:4 estimated:1 neuroscience:1 per:1 promise:1 group:1 key:1 four:3 rectangle:1 graph:1 run:1 schneiderman:1 cheek:1 decision:2 illuminated:1 bound:4 correspondence:2 quadratic:2 occur:1 scene:1 aspect:1 department:1 developing:1 according:1 combination:10 son:1 separability:1 cun:1 biologically:1 gradually:1 invariant:2 previously:1 turn:1 mechanism:1 bsat:1 nose:5 operation:2 hierarchical:3 robustness:3 ho:1 thomas:2 original:2 build:2 added:1 quantity:1 question:5 already:1 amongst:1 distance:2 separating:1 seven:1 collected:1 reason:1 o1:2 modeled:1 minimizing:2 pie:2 difficult:1 negative:2 memo:1 synthesizing:1 perform:2 upper:2 riesenhuber:1 viola:1 head:7 precise:1 y1:2 varied:1 arbitrary:2 intensity:1 bernd:1 pair:1 required:1 fort:1 suggested:1 usually:1 pattern:14 fp:1 challenge:2 eyebrow:1 video:1 mouth:5 power:1 natural:2 haar:1 indicator:1 eye:5 started:1 extract:2 siena:2 morphing:1 vancouver:1 determining:2 relative:2 interesting:1 versus:1 degree:2 viewpoint:1 face:45 taking:1 regard:1 curve:1 depth:3 calculated:1 transaction:1 uni:1 keep:1 global:7 pittsburgh:1 consuming:1 discriminative:5 knew:1 search:4 lip:1 kanade:2 learn:4 nature:2 expanding:1 expansion:3 papageorgiou:2 main:1 motivation:1 whole:4 allowed:1 body:1 fig:6 roc:1 wiley:1 position:8 extractor:1 wavelet:1 learns:2 down:1 specific:1 showing:1 svm:26 false:1 vapnik:1 phd:1 mohan:1 illumination:3 margin:2 boston:1 visual:2 satisfies:1 determines:1 extracted:4 ma:5 conditional:1 towards:1 considerable:1 change:4 feasible:1 determined:1 baluja:1 hyperplane:2 called:1 invariance:3 experimental:1 support:5 people:1 arises:1 collins:1 frontal:2 trainable:1 |
1,218 | 211 | A Large-Scale Neural Network
A LARGE-SCALE NEURAL NETWORK
WHICH RECOGNIZES HANDWRITTEN
KANJI CHARACTERS
Yoshihiro Mori
Kazuki Joe
ATR Auditory and Visual Perception Research Laboratories
Sanpeidani Inuidani Seika-cho Soraku-gun Kyoto 619-02 Japan
ABSTRACT
We propose a new way to construct a large-scale neural network for
3.000 handwritten Kanji characters recognition. This neural network
consists of 3 parts: a collection of small-scale networks which are
trained individually on a small number of Kanji characters; a network
which integrates the output from the small-scale networks, and a
process to facilitate the integration of these neworks. The recognition
rate of the total system is comparable with those of the small-scale
networks. Our results indicate that the proposed method is effective for
constructing a large-scale network without loss of recognition
performance.
1 INTRODUCTION
Neural networks have been applied to recognition tasks in many fields. with good results
[Denker, 1988][Mori,1988][Weideman, 1989]. They have performed better than
conventional methods. However these networks currently operate with only a few
categories, about 20 to 30. The Japanese writing system at present is composed of about
3,000 characters. For a network to recognize this many characters, it must be given a
large number of categories while maintaining its level of performance.
To train small-scale neural networks is not a difficult task. Therefore. exploring methods
for integrating these small-scale neural networks is important to construct a large-scale
network. If such methods could integrate small-scale networks without loss of the
performance, the scale of neural networks would be extended dramatically. In this paper,
we propose such a method for constructing a large-scale network whose object is to
recognize 3,000 handwritten Kanji characters, and report the result of a part of this
network. This method is not limited to systems for character recognition, and can be
applied to any system which recognizes many categories.
2 STRATEGIES FOR A LARGE-SCALE NETWORK
Knowing the current recognition and generalization capacity of a neural network. we
realized that constructing a large-scale monolithic network would not be efficient or
415
416
Mori and Joe
effective. Instead, from the start we decided on a building blocks approach
[Mori,1988] [Waibel,1988]. There are two strategies to mix many small-scale networks.
2.1 Selective Neural Network (SNN)
In this strategy, a large-scale neural network is made from many small-scale networks
which are trained individually on a small number of categories, and a network (SNN)
which selects the appropriate small-scale network (Fig. I). The advantage of this strategy
is that the information passed to a selected small-scale networks is always appropriate for
that network. Therefore, training these small-scale networks is very easy. But on the
other hand, increasing the number of categories will substantially increase the training
time of the SNN, and may make it harder for the SNN to retain high perfonnance.
Furthennore, the error rate of the SNN will limit the perfonnance of the whole system.
2.2 Integrative Neural Network (INN)
In this strategy, a large-scale neural network is made from many small-scale networks
which are trained individually on a small number of categories. and a network (INN)
which integrates the output from these small-scale networks(Fig. 2). The advantage of
this strategy is that every small-scale network gets information and contributes to finding
the right answer. Therefore, it is possible to use the knowledge distributed among each
small-scale network. But in some respects. various devices are needed to make the
integration easier.
The common advantage with both strategies just mentioned is that the size of each neural
network is relatively small, and it does not take a long time to train these networks. Each
small-scale networks is considered an independent part of the whole system. Therefore,
retraining these networks (to improve the performance of the whole system) will not take
too long.
~__....
O,utput
Sub
Net
1
? ?
Neural Network
(Selection Type)
':1U:U/W:::::::/:::/E::::::::.
: : Suspending
/ Network
(:::::{::::::::::::::::.:::::::: ~
Fig. 1 SNN Strategy
A Large-Scale Neural Network
Output
Neural Network
(Integration Type)
? ?
Fig. 2 INN Strategy
3 STRUCTURE OF LARGE-SCALE NETWORK
The whole system is constructed using three kinds of neural networks. The ftrst one,
called a SubNet, is an ordinary three layered feed forward type neural network trained
using the Back Propagation learning algorithm. The second kind of network is called a
SuperNet. This neural network makes its decision by integrating the outputs from all the
SubNets. This network is also a 3-layered feed-forward net, but is larger than the Subnets.
The last network, which we call an OtherFilter, is devised to improve the integration of
the S uperNet. This OtherFilter network was designed using the L VQ algorithm
[Khonen,1988]. There are also some changes made in the BP learning algorithm
especially for pattern recognition [Joe,1989].
We decided that, based on the time it takes for learning, there should be 9 categories in
each small-scale network. The 3,000 characters are separated into these small groups
through the K-means clustering method, which allows similar characters to be grouped
together. The separation occurs in two stages. First, 11 groups of 270 characters each are
formed, then each group is separated into 30 smaller units. In this way, 330 groups of 9
characters each are obtained. We choose the INN strategy to use distributed knowledge to
full advantage. The 9-character units are SubNets, which are integrated in 2 stages. First
30 SubNets are integrated by a higher level network SuperNet. Altogether, 11 SuperNets
are needed to recognize all 3,000 characters. SuperNets are in turn integrated by a higher
level network, the HyperNet. More precisely, the role and structure of these kinds of
networks are as follows:
3.1 SubNet
A feature vector extracted from handwritten patterns is used as the input (described in
Section 4.1). The number of units in the output layer is the same as the number of
categories to be recognized by the SubNet. In short, the role of a SubNet is to output the
similarity between the input pattern and the categories allotted to the SubNet. (Fig. 3)
3.2 SuperNet
The outputs from each SubNet fIltered by the OtherFilter network are used as the input to
417
418
Mori and Joe
the SuperNet. The number of units in an output layer is the same as the number of
SubNets belonging to a SuperNet. In shortt the role of SuperNet is to select the SubNet
which covers the category corresponding to the input patterns. (Fig. 5)
Output
Horizontal
+45?diagonal
Vertical
Original Pattern
Fig. 3 S ubNet
3.3 OtherFIIter
45(9x5) reference vectors are assigned to each SubNet. LVQ is used to adapt these
reference vectors t so that each input vector has a reference of the correct SubNet as its
closest reference vector. The
OtherFilter method is to frrst measure
the distance between all the reference
d
vectors and one input vector. The
mean distance and normal deviation of
distance are calculated. The distance
between a S ubNet and an input vector
is defmed to be the smallest distance
of that SubNet's reference vectors to
? References
the input vector .
?
?
XInput Vector
Fig4. Shape of OtherFilter
?
f(xo}=l 1(1+ e (x n-M+2d)/Cd) (1)
Xn : The Distance of Nth SubNet
M : The Mean of Xn
d : The Variance of Xn
C : Constant
A Large-Scale Neural Network
This distance modified by equation (1) is multiplied by the outputs of the SubNet. and fed
into the SuperNet. The outputs of SubNets whose distance is greater than the mean
distance are suppressed. and the outputs of SubNets whose distance is smaller than the
mean distance are amplified. In this way. the outputs of SubNets are modified to improve
the integration of the higher level SuperNet. (Fig. 5)
HyperNet 1
SuperNet 11
SubNet 330
OtherFilter 12
Other-Filter
FigS. Outline of the Whole System
4 RECOGNITION EXPERIMENT
4.1 TRAINING PATTERN
The training samples for this network were chosen from a database of about 3000 Kanji
characters [Saito 1985]. For each character. there are 200 handwritten samples from
different writers. 100 are used as training samples. and the remaining 100 are used to test
recognition accuracy of the trained network. All samples in the database consist of 64 by
63 dots.
419
420
Mori and Joe
JlQ
~
~~
~
.J-~
~~lJ
~~
~ ~
~
,~
~~
~
~~
-V'#f)
~
~
.orfffi
~~
J..~
~i2
O~
~
DI~J
o/N{
Fig. 6 Examples of training pattern
4.2 LDCD FEATURE
If we were to use this pattern as the input to our neural net, the number of units required
in the input layer would be too large for the computational abilities of current computers.
Therefore, a feature vector extracted from the handwritten patterns is used as the input. In
the "LDCD feature" [Hagita 1983], there are 256 dimensions computing a line segment
length along four directions: horizontal, vertical, and two diagonals in the 8 by 8 squares
into which the handwritten samples are divided.
t"
:61
o
horizontal
component
Fig 7. LDCD Feature
4.3 RECOGNITION RESULTS
In the work reported here, one SuperNet, 30 SubNets and one OtherFilter were
constructed for recognition experiments. SubNets were trained until the recognition of
training samples reaches at least 99%. With these SubNets, the mean recognition rate of
test patterns was 92%. This recognition rate is higher than that of conventional methods.
A SuperNet which integrates the output modified by OtherFilter from 30 trained SubNets
A Large-Scale Neural Network
was then constructed. The number of units in the input layer of the SuperNet was 270.
This SuperNet was trained until the performance of training samples becomes at least
93%. With this SuperNet, the recognition rate of test patterns was 74%, though that of
OtherFilter was 72%. The recognition rate of a system without the OtherFilter of test
patterns was 55%.
5 CONCLUSION
We have here proposed a new way of constructing a large-scale neural network for the
recognition of 3,000 handwritten Kanji characters. With this method, a system
recognizing 270 Kanji characters was constructed. This system will become a part of a
system recognizing 3,000 Kanji characters. Only a modest training time was necessary
owing to the modular nature of the system. Moreover, this modularity means that only a
modest re-training time is necessary for retraining an erroneous neural network in the
whole system. The overall system performance can be improved by retraining just that
neural network, and there is no need to retrain the whole system. However, the
performance of the OtherFilter is not satisfactory. We intend to improve the OtherFilter,
and build a large-scale network for the recognition of 3,000 handwritten Kanji characters
by the method reported here.
Acknowledgments
We are grateful to Dr. Yodogawa for his support and encouragement. Special thanks to
Dr. Sei Miyake for the ideas he provided in our many discussions. The authors would like
to acknowledge, with thanks, the help of Erik McDermott for his valuable assistance in
writing this paper in English.
References
[Denier, 1988]
l.S.Denker, W.R.Gardner, H.P. Graf, D.Henderson, R.E. Howard,
W.Hubbard, L.DJackel. H.S.Baird, I.Guyon : "Neural Network Recognizer for HandWritten ZIP Code Digits", NEURAL INFORMATION PROCESSING SYSTEMS 1.
pp.323-331, Morgan Kaufmann. 1988
[Mori,1988]
Y.Mori. K.Yokosawa : "Neural Networks that Learn to Discriminate
Similar Kanji Characters". NEURAL INFORMATION PROCESSING SYSTEMS 1,
pp.332-339, Morgan Kaufmann. 1988
[Weideman.1989]W.E.Weideman. M.T.Manry. H.C.Yau ; tI A COMPARISON OF A
NEAREST NEIGHBOR CLASSIFIER AND A NEURAL NETWORK FOR NUMERIC
HANDPRINT CHARACTER RECOGNITION". UCNN89(Washington), VoLl, pp.117120, June 1989
421
422
Mori and Joe
Alex Waibel, "Consonant Recognition by Modular Construction of
[Waibel, 1988]
Large Phonemic Time-Delay Neural Networks", NEURAL INFORMATION
PROCESSING SYSTEMS 1, pp.215-223, Morgan Kaufmann, 1988
[Joo,1989]
KJoo, Y.Mori, S.Miyake : "Simulation of a Large-Scale Neural
Networks on a Parallel Computer", 4th Hypercube Concurrent Computers,1989
[Khonen,1988] T.Kohonen, G.Barna, R.Chrisley : "Statistical Pattern Recognition
with Neural Networks", IEEE, Proc.of ICNN, YoU, pp.61-68, July 1988
[Saito,1985]
T.Saito, H.Yamada, K.Yamamoto : "On the Data Base ETL9 of
Handprinted Characters in 1IS Chinese Characters and Its Analysis", J68-D, 4, 757-764,
1985
[Hagita,1983]
N.Hagita, S.Naito, I.Masuda : "Recognition of Handprinted Chinese
Characters by Global and Local Direction Contributivity Density-Feature", J66-D, 6,
722-729,1983
| 211 |@word especially:1 build:1 chinese:2 hypercube:1 indicate:1 sanpeidani:1 retraining:3 direction:2 assigned:1 intend:1 correct:1 integrative:1 filter:1 simulation:1 owing:1 laboratory:1 satisfactory:1 i2:1 x5:1 assistance:1 strategy:10 defmed:1 diagonal:2 distance:11 harder:1 subnet:13 atr:1 capacity:1 generalization:1 icnn:1 gun:1 neworks:1 outline:1 exploring:1 erik:1 current:2 length:1 code:1 considered:1 normal:1 must:1 handprinted:2 common:1 difficult:1 ftrst:1 shape:1 smallest:1 hypernet:2 designed:1 recognizer:1 proc:1 integrates:3 he:1 selected:1 device:1 currently:1 handprint:1 sei:1 vertical:2 individually:3 manry:1 grouped:1 hagita:3 short:1 yamada:1 hubbard:1 concurrent:1 filtered:1 encouragement:1 howard:1 acknowledge:1 extended:1 always:1 dot:1 modified:3 joo:1 similarity:1 along:1 constructed:4 base:1 become:1 closest:1 consists:1 required:1 june:1 inuidani:1 barna:1 seika:1 mcdermott:1 morgan:3 greater:1 integrated:3 lj:1 zip:1 perception:1 snn:6 recognized:1 pattern:13 increasing:1 becomes:1 provided:1 selective:1 moreover:1 selects:1 overall:1 among:1 july:1 full:1 facilitate:1 mix:1 kind:3 kyoto:1 substantially:1 adapt:1 integration:5 special:1 long:2 nth:1 field:1 finding:1 construct:2 devised:1 washington:1 divided:1 improve:4 every:1 gardner:1 ti:1 classifier:1 report:1 denier:1 unit:6 few:1 shortt:1 composed:1 recognize:3 graf:1 monolithic:1 yodogawa:1 local:1 limit:1 loss:2 extracted:2 operate:1 integrate:1 henderson:1 call:1 limited:1 cd:1 easy:1 realized:1 decided:2 acknowledgment:1 occurs:1 last:1 english:1 block:1 necessary:2 idea:1 knowing:1 digit:1 perfonnance:2 modest:2 yamamoto:1 neighbor:1 saito:3 re:1 ldcd:3 distributed:2 calculated:1 passed:1 xn:3 dimension:1 integrating:2 numeric:1 forward:2 collection:1 made:3 get:1 cover:1 soraku:1 selection:1 layered:2 author:1 ordinary:1 dramatically:1 writing:2 deviation:1 conventional:2 recognizing:2 delay:1 global:1 too:2 reported:2 miyake:2 category:10 answer:1 consonant:1 masuda:1 cho:1 thanks:2 density:1 modularity:1 his:2 retain:1 chrisley:1 nature:1 learn:1 erroneous:1 together:1 contributes:1 construction:1 group:4 four:1 voll:1 japanese:1 utput:1 constructing:4 choose:1 dr:2 recognition:22 yau:1 whole:7 japan:1 database:2 fig4:1 role:3 you:1 fig:11 retrain:1 baird:1 guyon:1 separation:1 sub:1 performed:1 decision:1 valuable:1 mentioned:1 comparable:1 start:1 layer:4 parallel:1 kazuki:1 kanji:10 trained:8 grateful:1 naito:1 segment:1 formed:1 accuracy:1 square:1 variance:1 kaufmann:3 writer:1 precisely:1 alex:1 bp:1 consist:1 joe:6 handwritten:10 various:1 weideman:3 train:2 separated:2 relatively:1 effective:2 waibel:3 easier:1 reach:1 belonging:1 smaller:2 whose:3 modular:2 larger:1 character:24 suppressed:1 visual:1 pp:5 furthennore:1 yokosawa:1 ability:1 di:1 xo:1 auditory:1 mori:10 advantage:4 equation:1 knowledge:2 net:3 inn:4 vq:1 yoshihiro:1 propose:2 turn:1 needed:2 lvq:1 back:1 kohonen:1 fed:1 feed:2 change:1 higher:4 ubnet:2 multiplied:1 improved:1 amplified:1 denker:2 appropriate:2 frrst:1 though:1 total:1 called:2 just:2 stage:2 discriminate:1 until:2 hand:1 altogether:1 horizontal:3 original:1 select:1 allotted:1 clustering:1 remaining:1 propagation:1 object:1 help:1 recognizes:2 supernet:14 maintaining:1 subnets:12 support:1 nearest:1 phonemic:1 building:1 |
1,219 | 2,110 | A Neural Oscillator Model of Auditory
Selective Attention
Stuart N. Wrigley and Guy J. Brown
Department of Computer Science, University of Sheffield, Regent Court,
211 Portobello Street, Sheffield S1 4DP, UK.
[email protected], [email protected]
Abstract
A model of auditory grouping is described in which auditory attention
plays a key role. The model is based upon an oscillatory correlation
framework, in which neural oscillators representing a single perceptual
stream are synchronised, and are desynchronised from oscillators
representing other streams. The model suggests a mechanism by which
attention can be directed to the high or low tones in a repeating sequence
of tones with alternating frequencies. In addition, it simulates the
perceptual segregation of a mistuned harmonic from a complex tone.
1
Introduction
In virtually all listening situations, we are exposed to a mixture of sound energy from
multiple sources. Hence, the auditory system must separate an acoustic mixture in order to
create a perceptual description of each sound source. It has been proposed that this process
of auditory scene analysis (ASA) [2] takes place in two conceptual stages: segmentation
in which the acoustic mixture is separated into its constituent ?atomic? units, followed by
grouping in which units that are likely to have arisen from the same source are
recombined. The perceptual ?object? produced by auditory grouping is called a stream.
Each stream describes a single sound source.
Few studies have investigated the role of attention in ASA; typically, ASA is seen as a
precursor to attentional mechanisms, which simply select one stream as the attentional
focus. Recently, however, it has been suggested that attention plays a much more
prominent role in ASA. Carlyon et al. [4] investigated how attention influences auditory
grouping with the use of a rapidly repeating sequence of high and low tones. It is known
that high frequency separations and/or high presentation rates encourage the high tones
and low tones to form separate streams, a phenomenon known as auditory streaming [2].
Carlyon et al. demonstrated that auditory streaming did not occur when listeners attended
to an alternative stimulus presented simultaneously. However, when they were instructed
to attend to the tone sequence, auditory streaming occurred as normal. From this, it was
concluded that attention is required for stream formation and not only for stream selection.
It has been proposed that attention can be divided into two different levels [9]: low-level
exogenous attention which groups acoustic elements to form streams, and a higher-level
endogenous mechanism which performs stream selection. Exogenous attention may overrule conscious (endogenous) selection (e.g. in response to a sudden loud bang). The work
presented here incorporates these two types of attention into a model of auditory grouping
(Figure 1). The model is based upon the oscillatory correlation theory [10], which
suggests that neural oscillations encode auditory grouping. Oscillators corresponding to
grouped auditory elements are synchronised, and are desynchronised from oscillators
encoding other groups. This theory is supported by neurobiological findings that report
ALI
Correlogram
Signal
Cochlear
Filtering
Hair
cell
Cross
Channel
Correlation
Attentional
Stream
Neural
Oscillator
Network
Figure 1: Schematic diagram of the model (the attentional leaky integrator is labelled ALI).
synchronised oscillations in the auditory system [6]. Within the oscillatory correlation
framework, attentional selection can be implemented by synchronising attentional activity
with the stream of interest.
2
The model
2.1 Auditory periphery
Cochlear filtering is modelled by a bank of 128 gammatone filters with centre frequencies
equally spaced on the equivalent rectangular bandwidth (ERB) scale between 50 Hz and 2.5
kHz [3]. Auditory nerve firing rate is approximated by half-wave rectifying and square root
compressing the output of each filter. Input to the model is sampled at a rate of 8 kHz.
2.2 Pitch and harmonicity analysis
It is known that a difference in fundamental frequency (F0) can assist the perceptual
segregation of complex sounds [2]. Accordingly, the second stage of the model extracts pitch
information from the simulated auditory nerve responses. This is achieved by computing the
autocorrelation of the activity in each channel to form a correlogram [3]. At time t, the
autocorrelation of channel i with lag ? is given by:
P ?1
A ( i, t , ? ) =
? r ( i, t ? k )r ( i, t ? k ? ? )w ( k )
(1)
k=0
Here, r is the auditory nerve activity. The autocorrelation for channel i is computed using a 25
ms rectangular window w (P = 200) with lag steps equal to the sampling period, up to a
maximum lag of 20 ms. When the correlogram is summed across frequency, the resulting
?summary correlogram? exhibits a large peak at the lag corresponding to the fundamental
period of the stimulus. An accurate estimate of the F0 is found by fitting a parabolic curve to
the three samples centred on the summary peak.
The correlogram may also be used to identify formant and harmonic regions due to their
similar patterns of periodicity [11]. This is achieved by computing the correlations between
adjacent channels of the correlogram as follows:
L ?1
1
C ( i ) = --L
? A? ( i, t, ? )A? ( i + 1, t, ? )
(2)
?=0
Here, A?( i, t, ? ) is the autocorrelation function of (1) which has been normalised to have zero
mean and unity variance; L is the maximum autocorrelation lag in samples (L = 160).
2.3 Neural oscillator network
The network consists of 128 oscillators and is based upon the two-dimensional locally
excitatory globally inhibitory oscillator network (LEGION) of Wang [10], [11]. Within
LEGION, oscillators are synchronised by placing local excitatory links between them.
Additionally, a global inhibitor receives excitation from each oscillator, and inhibits every
oscillator in the network. This ensures that only one block of synchronised oscillators can be
active at any one time. Hence, separate blocks of synchronised oscillators - which correspond
to the notion of a segment in ASA - arise through the action of local excitation and global
inhibition.
The model described here differs from Wang?s approach [10] in three respects. Firstly, the
network is one-dimensional rather than two-dimensional; we argue that this is more plausible.
Secondly, excitatory links can be global as well as local; this allows harmonically-related
segments to be grouped. Finally, we introduce an attentional leaky integrator (ALI), which
selects one block of oscillators to become the attentional stream (i.e., the stream which is in the
attentional foreground).
The building block of the network is a single oscillator, which consists of a reciprocally
connected excitatory unit and inhibitory unit whose activities are represented by x and y
respectively:
3
x? = 3x ? x + 2 ? y + I o
(3a)
x
y? = ? ? ? 1 + tanh ---? ? y
?
??
(3b)
Here, ?, ? and ? are parameters. Oscillations are stimulus dependent; they are only observed
when Io > 0, which corresponds to a periodic solution to (3) in which the oscillator cycles
between an ?active? phase and a ?silent? phase. The system may be regarded as a model for the
behaviour of a single neuron, or as a mean field approximation to a group of connected
neurons. The input Io to oscillator i is a combination of three factors: external input Ir ,
network activity and global inhibition as follows:
Io = I r ?W z S ( z, ? z ) +
? Wik S ( xk, ?x )
(4)
k?i
Here, Wik is the connection strength between oscillators i and k; xk is the activity of oscillator
k. The parameter ?x is a threshold above which an oscillator can affect others in the network
and Wz is the weight of inhibition from the global inhibitor z. Similar to ?x, ?z acts a threshold
above which the global inhibitor can affect an oscillator. S is a squashing function which
compresses oscillator activity to be within a certain range:
1
S ( n, ? ) = -----------------------------?K ( n ? ? )
1+e
(5)
Here, K determines the sharpness of the sigmoidal function. The activity of the global inhibitor
is defined as
?
?
z? = H ? ? S ( xk, ? x ) ? 0.1? ? z
?
?
(6)
k
where H is the Heaviside function (H(n) = 1 for n ? 0, zero otherwise).
2.3.1 Segmentation
A block of channels are deemed to constitute a segment if the cross-channel correlation (2) is
greater than 0.3 for every channel in the block. Cross-correlations are weighted by the energy
of each channel in order to increase the contrast between spectral peaks and spectral dips.
These segments are encoded by a binary mask, which is unity when a channel contributes to a
segment and zero otherwise. To improve the resolution and separation of adjacent segments,
the cross-frequency spread of a segment is restricted to 3 channels. Oscillators within a
segment are synchronized by excitatory connections. The external input (Ir) of an oscillator
whose channel is a member of a segment is set to Ihigh otherwise it is set to Ilow.
2.3.2 Harmonicity grouping
Excitatory connections are made between segments if they are consistent with the current F0
estimate. A segment is classed as consistent with the F0 if a majority of its corresponding
correlogram channels exhibit a significant peak at the fundamental period (ratio of peak height
to channel energy greater than 0.46). A single connection is made between the centres of
harmonically related segments subject to old-plus-new constraints.
The old-plus-new heuristic [2] refers to the auditory system?s preference to ?interpret any part
of a current group of acoustic components as a continuation of a sound that just occurred? .
This is incorporated into the model by attaching ?age trackers? to each channel of the network.
Excitatory links are placed between harmonically related segments only if the two segments
are of similar age. The age trackers are leaky integrators:
+
B? k = d ( g [ M k ? B k ] ? [ 1 ? H ( M k ? Bk ) ]cBk )
(7)
= 0 otherwise. Mk is the (binary) value of the segment mask at
Here, [n] = n if n ? 0 and
channel k; small values of c and d result in a slow rise (d) and slow decay (c) for the integrator.
g is a gain factor.
+
[n]+
Consider two segments that start at the same time; the age trackers for their constituent
channels receive the same input, so the values of Bk will be the same. However, if two
segments start at different times, the age trackers for the earlier segment will have already
increased to a non-zero value when the second segment starts. This ?age difference? will
dissipate over time, as the values of both sets of leaky integrators approach unity.
2.3.3 Attentional leaky integrator (ALI)
Each oscillator is connected to the attentional leaky integrator (ALI) by excitatory links; the
strength of these connections is modulated by endogenous attention. Input to the ALI is given
by:
?
?
?
ali = H ? ? S ( x k, ? x )T k ? ? ALI? ? ali
?
?
(8)
k
?ALI is a threshold above which network activity can influence the ALI. Tk is an attentional
weighting which is related to the endogenous interest at frequency k:
T k = 1 ? ( 1 ? A k )L
(9)
Here, Ak is the endogenous interest at frequency k and L is the leaky integrator defined as:
+
L? = a ( b [ R ? L ] ? [ 1 ? H ( R ? L ) ]fL )
(10)
Small values of f and a result in a slow rise (a) and slow decay (f) for the integrator. b is a gain
factor. R = H ( xmax ) where xmax is the largest output activity of the network. The build-up of
attentional interest is therefore stimulus dependent. The attentional interest itself is modelled
as a Gaussian according to the gradient model of attention [7]:
A k = max A e
k
k?p
?----------22?
(11)
Here, Ak is the normalised attentional interest at frequency channel k and maxAk is the
maximum value that Ak can attain. p is the channel at which the peak of attentional interest
occurs, and ? determines the width of the peak.
A segment or group of segments are said to be attended to if their oscillatory activity coincides
temporally with a peak in the ALI activity. Initially, the connection weights between the
oscillator array and the ALI are strong: all segments feed excitation to the ALI, so all segments
are attended to. During sustained activity, these weights relax toward the Ak interest vector
such that strong weights exist for channels of high attentional interest and low weights exist
for channels of low attentional interest. ALI activity will only coincide with activity of the
channels within the attentional interest peak and any harmonically related (synchronised)
activity outside the Ak peak. All other activity will occur within a trough of ALI activity. This
behaviour allows both individual tones and harmonic complexes to be attended to using only a
single Ak peak.
The parameters for all simulations reported here were ? = 0.4, ? = 6.0, ? = 0.1, Wz = 0.5, ?z =
0.1, ?x = -0.5 and K = 50, d = 0.001, c = 5, g = 3, a = 0.0005, f = 5, b = 3, maxAk = 1, ? = 3, ?ALI
= 1.5, Ilow = -5.0, Ihigh = 0.2.The inter- and intra- segment connections have equal weights of
1.1.
3
Evaluation
Throughout this section, output from the model is represented by a ?pseudospectrogram? with
time on the abscissa and frequency channel on the ordinate. Three types of information are
superimposed on each plot. A gray pixel indicates the presence of a segment at a particular
frequency channel, which is also equivalent to the external input to the corresponding
oscillator: gray signifies Ihigh (causing the oscillator to be stimulated) and white signifies Ilow
(causing the oscillator to be unstimulated). Black pixels represent active oscillators (i.e.
oscillators whose x value exceeds a threshold value). At the top of each figure, ALI activity is
shown. Any oscillators which are temporally synchronised with the ALI are considered to be
in the attentional foreground.
3.1 Segregation of a component from a harmonic complex
Darwin et al. [5] investigated the effect of a mistuned harmonic upon the pitch of a 12
component complex tone. As the degree of mistuning of the fourth harmonic increased
towards 4%, the shift in the perceived pitch of the complex also increased. This effect was less
pronounced for mistunings of more than 4%; beyond 8% mistuning, little pitch shift was
observed. Apparently, the pitch of a complex tone is calculated using only those channels
which belong to the corresponding stream. When the harmonic is subject to mistunings below
8%, it is grouped with the rest of the complex and so can affect the pitch percept. Mistunings
of greater than 8% cause the harmonic to be segregated into a second stream, and so it is
excluded from the pitch percept.
B
C
1.5
Pitch shift (Hz)
120
100
120
100
80
60
40
20
Channel
Channel
Channel
A
120
100
80
60
40
20
80
60
40
20
D
Darwin
Model
1
0.5
0
0
20
40 60
Time (ms)
80
0
20
40 60
Time (ms)
80
0
20
40 60
Time (ms)
80
0
2
4
6
8
Mistuning of 4th harmonic (%)
Figure 2: A,B,C: Network response to mistuning of the fourth harmonic of a 12 harmonic
complex (0%, 6% and 8% respectively). ALI activity is shown at the top of each plot.
Gray areas denote the presence of a segment and black areas denote oscillators in the
active phase. Arrows show the focus of attentional interest. D: Pitch shift versus degree of
mistuning. A Gaussian derivative is fitted to each data set.
120
Channel
100
80
60
40
20
0
100
200
300
Time (ms)
400
500
600
Figure 3: Captor tones preceding the complex capture the fourth harmonic into a separate
stream. ALI activity (top) shows that this harmonic is the focus of attention and would be
?heard out? . The attentional interest vector (Ak) is shown to the right of the figure.
This behaviour is reproduced by our model (Figure 2). All the oscillators at frequency
channels corresponding to harmonics are temporally synchronised for mistunings up to 8%
(plots A and B) signifying that the harmonics belong to the same perceptual group. Mistunings
beyond 8% cause the mistuned harmonic to become desychronised from the rest of the
complex (plot C) - two distinct perceptual groups are now present: one containing the fourth
harmonic and the other containing the remainder of the complex tone. A comparison of the
pitch shifts found by Darwin et al. and the shifts predicted by the model is shown in plot D.
The pitch of the complex was calculated by creating a summary correlogram (similar to that
used in section 2.2) using frequency channels contained within the complex tone group. Only
segment channels below 1.1 kHz were used for this summary since low frequency (resolved)
harmonics are known to dominate the pitch percept [8].
Darwin et al. also showed that the effect of mistuning was diminished when the fourth
harmonic was ?captured? from the complex by four preceding tones at the same frequency. In
this situation, no matter how small the mistuning, the harmonic is segregated from the
complex and does not influence the pitch percept. Figure 3 shows the capture of the harmonic
with no mistuning. Attentional interest is focused on the fourth harmonic: oscillator activity
for the captor tone segments is synchronised with the ALI activity. During the 550 ms before
the complex tone onset, the age tracker activities for the captor tone channels build up. When
the complex tone begins, there is a significant age difference between the frequency channels
stimulated by the fourth harmonic and those stimulated by the remainder of the complex. Such
a difference prevents excitatory harmonicity connections from being made between the fourth
harmonic and the remaining harmonics. This behaviour is consistent with the old-plus-new
heuristic; a current acoustic event is interpreted as a continuation of a previous stimulus.
The old-plus-new heuristic can be further demonstrated by starting the fourth harmonic before
the rest of the complex. Figure 4 shows the output of the model when the fourth harmonic is
subject to a 50 ms onset asynchrony. During this time, the age trackers of channels excited by
the fourth harmonic increase to a significantly higher value than those of the remaining
harmonics. Once again, this prevents excitatory connections being made between the fourth
harmonic and the other harmonically related segments. The early harmonic is desynchronised
from the rest of the complex: two streams are formed. However, after a period of time, the
importance of the onset asynchrony decreases as the channel ages approach their maximal
values. Once this occurs, there is no longer any evidence to prevent excitatory links from
being made between the fourth harmonic and the rest of the complex. Grouping by
harmonicity then occurs for all segments: the complex and the early harmonic synchronise to
form a single stream.
3.2 Auditory streaming
Within the framework presented here, auditory streaming is an emergent property; all events
which occur over time, and are subject to attentional interest, are implicitly grouped. Two
temporally separated events at different frequencies must both fall under the Ak peak to be
grouped. It is the width of the Ak peak that determines frequency separation-dependent
streaming, rather than local connections between oscillators as in [10]. The build-up of
streaming [1] is modelled by the leaky integrator in (9). Figure 5 shows the effect of two
different frequency separations on the ability of the network to perform auditory streaming and
shows a good match to experimental findings [1], [4]. At low frequency separations, both the
high and low frequency segments fall under the attentional interest peak; this allows the
oscillator activities of both frequency bands to influence the ALI and hence they are
considered to be in the attentional foreground. At higher frequency separations, one of the
frequency bands falls outside of the attentional peak (in this example, the high frequency tones
fall outside) and hence it cannot influence the ALI. Such behaviour is not seen immediately,
because the attentional interest vector is subject to a build up effect as described in (9).
Initially the attentional interest is maximal across all frequencies; as the leaky integrator value
increases, the interest peak begins to dominate and interest in other frequencies tends toward
zero.
4
Discussion
A model of auditory attention has been presented which is based on previous neural oscillator
work by Wang and colleagues [10], [11] but differs in two important respects. Firstly, our
network is unidimensional; in contrast, Wang?s approach employs a two-dimensional timefrequency grid for which there is weak physiological justification. Secondly, our model
regards attention as a key factor in the stream segregation process. In our model, attentional
interest may be consciously directed toward a particular stream, causing that stream to be
selected as the attentional foreground.
Few auditory models have incorporated attentional effects in a plausible manner. For example,
Wang?s ?shifting synchronisation? theory [3] suggests that attention is directed towards a
stream when its constituent oscillators reach the active phase. This contradicts experimental
findings, which suggest that attention selects a single stream whose salience is increased for a
sustained period of time [2]. Additionally, Wang?s model fails to account for exogenous
reorientation of attention to a sudden loud stimulus; the shifting synchronisation approach
would multiplex it as normal with no attentional emphasis. By ensuring that the minimum Ak
value for the attentional interest is always non-zero, it is possible to weight activity outside of
the attentional interest peak and force it to influence the ALI. Such weighting could be derived
from a measure of the sound intensity present in each frequency channel.
We have demonstrated the model? s ability to accurately simulate a number of perceptual
phenomena. The time course of perception is well simulated, showing how factors such as
mistuning and onset asynchrony can cause a harmonic to be segregated from a complex tone.
It is interesting to note that a good match to Darwin?s pitch shift data (Figure 2D) was only
found when harmonically related segments below 1.1 kHz were used. The dominance of lower
(resolved) harmonics on pitch is well known [8], and our findings suggest that the correlogram
does not accurately model this aspect of pitch perception.
120
Channel
100
80
60
40
20
0
50
100
150
200
250
300
350
Time (ms)
Figure 4: Asynchronous onset of the fourth harmonic causes it to segregate into a separate
stream. The attentional interest vector (Ak) is shown to the right of the figure.
100
Channel
Channel
100
90
80
90
80
0
200
400
Time (ms)
600
0
200
400
Time (ms)
600
Figure 5: Auditory streaming at frequency separations of 5 semitones (left) and 3
semitones (right). Streaming occurs at the higher separation. The timescale of adaptation
for the attentional interest has been reduced to aid the clarity of the figures.
The simulation of two tone streaming shows how the proposed attentional mechanism and its
cross-frequency spread accounts for grouping of sequential events according to their proximity
in frequency. A sequence of two tones will only stream if one set of tones fall outside of the
peak of attentional interest. Frequency separations for streaming to occur in the model (greater
than 3 to 4 semitones) are in agreement with experimental data, as is the timescale for the
build-up of the streaming effect [1].
In summary, we have proposed a physiologically plausible model in which auditory streams
are encoded by a unidimensional neural oscillator network. The network creates auditory
streams according to grouping factors such as harmonicity, frequency proximity and common
onset, and selects one stream as the attentional foreground. Current work is concentrating on
expanding the system to include binaural effects, such as inter-ear attentional competition [4].
References
[1] Anstis, S. & Saida, S. (1985) Adaptation to auditory streaming of frequency-modulated tones. J. Exp.
Psychol. Human 11 257-271.
[2] Bregman, A. S. (1990) Auditory Scene Analysis. Cambridge MA: MIT Press.
[3] Brown, G. J. & Cooke, M. (1994) Computational auditory scene analysis. Comput. Speech Lang. 8,
pp. 297-336.
[4] Carlyon, R. P., Cusack, R., Foxton, J. M. & Robertson, I. H. (2001) Effects of attention and unilateral
neglect on auditory stream segregation. J. Exp. Psychol. Human 27(1) 115-127.
[5] Darwin, C. J., Hukin, R. W. & Al-Khatib, B. Y. (1995) Grouping in pitch perception: Evidence for
sequential constraints. J. Acoust. Soc. Am. 98(2)Pt1 880-885.
[6] Joliot, M., Ribary, U. & Llin?s, R. (1994) Human oscillatory brain activity near 40 Hz coexists with
cognitive temporal binding. Proc. Natl. Acad. Sci. USA 91 11748-51.
[7] Mondor, T. A. & Bregman, A. S. (1994) Allocating attention to frequency regions. Percept.
Psychophys. 56(3) 268-276.
[8] Moore, B. C. J. (1997) An introduction to the psychology of hearing. Academic Press.
[9] Spence, C. J., Driver, J. (1994) Covert spatial orienting in audition: exogenous and endogenous
mechanisms. J. Exp. Psychol. Human 20(3) 555-574.
[10] Wang, D. L. (1996) Primitive auditory segregation based on oscillatory correlation. Cognitive Sci.
20 409-456.
[11] Wang, D. L. & Brown, G. J. (1999) Separation of speech from interfering sounds based on
oscillatory correlation. IEEE Trans. Neural Networks 10 684-697.
| 2110 |@word timefrequency:1 simulation:2 excited:1 attended:4 current:4 lang:1 must:2 plot:5 half:1 selected:1 tone:24 accordingly:1 xk:3 sudden:2 preference:1 firstly:2 sigmoidal:1 height:1 harmonically:6 become:2 driver:1 consists:2 regent:1 sustained:2 fitting:1 autocorrelation:5 manner:1 introduce:1 inter:2 mask:2 abscissa:1 integrator:11 brain:1 globally:1 little:1 precursor:1 window:1 begin:2 interpreted:1 acoust:1 finding:4 synchronisation:2 temporal:1 every:2 act:1 uk:3 unit:4 before:2 attend:1 local:4 multiplex:1 tends:1 io:3 acad:1 encoding:1 ak:11 firing:1 black:2 plus:4 emphasis:1 suggests:3 range:1 directed:3 spence:1 atomic:1 block:6 differs:2 area:2 attain:1 significantly:1 refers:1 suggest:2 cannot:1 desynchronised:3 selection:4 influence:6 equivalent:2 demonstrated:3 primitive:1 attention:21 starting:1 rectangular:2 sharpness:1 resolution:1 focused:1 immediately:1 array:1 regarded:1 dominate:2 notion:1 justification:1 play:2 agreement:1 element:2 robertson:1 approximated:1 observed:2 role:3 wang:8 capture:2 region:2 compressing:1 ensures:1 connected:3 cycle:1 decrease:1 segment:32 exposed:1 asa:5 ali:25 upon:4 creates:1 binaural:1 resolved:2 emergent:1 represented:2 listener:1 separated:2 distinct:1 formation:1 outside:5 whose:4 lag:5 encoded:2 plausible:3 heuristic:3 pt1:1 relax:1 otherwise:4 ability:2 formant:1 timescale:2 itself:1 reproduced:1 sequence:4 maximal:2 remainder:2 adaptation:2 causing:3 rapidly:1 gammatone:1 description:1 pronounced:1 competition:1 constituent:3 object:1 tk:1 ac:2 strong:2 soc:1 implemented:1 predicted:1 synchronized:1 filter:2 human:4 behaviour:5 recombined:1 secondly:2 proximity:2 tracker:6 considered:2 normal:2 exp:3 early:2 perceived:1 proc:1 tanh:1 grouped:5 largest:1 create:1 weighted:1 mit:1 inhibitor:4 gaussian:2 always:1 rather:2 encode:1 derived:1 focus:3 superimposed:1 indicates:1 contrast:2 am:1 dependent:3 streaming:14 typically:1 initially:2 selective:1 selects:3 pixel:2 spatial:1 summed:1 equal:2 field:1 once:2 legion:2 sampling:1 placing:1 stuart:1 unstimulated:1 foreground:5 report:1 stimulus:6 others:1 few:2 employ:1 simultaneously:1 individual:1 joliot:1 phase:4 mistuned:3 interest:26 intra:1 evaluation:1 mixture:3 natl:1 accurate:1 allocating:1 bregman:2 encourage:1 old:4 darwin:6 mk:1 fitted:1 increased:4 earlier:1 signifies:2 hearing:1 reported:1 periodic:1 fundamental:3 peak:18 again:1 ear:1 containing:2 guy:1 external:3 creating:1 cognitive:2 derivative:1 audition:1 account:2 centred:1 matter:1 trough:1 onset:6 stream:30 dissipate:1 root:1 endogenous:6 exogenous:4 apparently:1 wave:1 start:3 rectifying:1 square:1 ir:2 formed:1 variance:1 percept:5 spaced:1 identify:1 correspond:1 modelled:3 weak:1 consciously:1 accurately:2 produced:1 oscillatory:7 reach:1 energy:3 colleague:1 frequency:35 pp:1 sampled:1 auditory:32 gain:2 concentrating:1 segmentation:2 nerve:3 feed:1 higher:4 response:3 just:1 stage:2 correlation:9 receives:1 gray:3 asynchrony:3 orienting:1 usa:1 effect:9 building:1 brown:4 hence:4 alternating:1 excluded:1 moore:1 overrule:1 white:1 adjacent:2 semitone:3 during:3 width:2 portobello:1 excitation:3 coincides:1 m:11 prominent:1 performs:1 covert:1 harmonic:36 recently:1 common:1 khz:4 belong:2 occurred:2 interpret:1 significant:2 cambridge:1 grid:1 centre:2 f0:4 longer:1 inhibition:3 showed:1 periphery:1 certain:1 binary:2 seen:2 minimum:1 captured:1 greater:4 preceding:2 period:5 signal:1 multiple:1 sound:7 exceeds:1 match:2 academic:1 cross:5 divided:1 equally:1 schematic:1 pitch:18 ensuring:1 sheffield:2 hair:1 represent:1 arisen:1 xmax:2 achieved:2 cell:1 receive:1 addition:1 shef:2 diagram:1 source:4 concluded:1 rest:5 hz:3 subject:5 virtually:1 simulates:1 member:1 incorporates:1 near:1 presence:2 affect:3 psychology:1 bandwidth:1 silent:1 unidimensional:2 court:1 listening:1 shift:7 khatib:1 assist:1 unilateral:1 speech:2 cause:4 constitute:1 action:1 heard:1 repeating:2 conscious:1 locally:1 band:2 mistuning:9 reduced:1 continuation:2 exist:2 inhibitory:2 ilow:3 group:8 key:2 erb:1 four:1 threshold:4 dominance:1 clarity:1 prevent:1 fourth:14 harmonicity:5 place:1 parabolic:1 throughout:1 separation:10 oscillation:3 fl:1 captor:3 followed:1 activity:27 strength:2 occur:4 constraint:2 scene:3 aspect:1 simulate:1 llin:1 inhibits:1 department:1 according:3 combination:1 describes:1 across:2 contradicts:1 unity:3 s1:1 restricted:1 segregation:6 mechanism:5 cbk:1 spectral:2 alternative:1 compress:1 top:3 remaining:2 include:1 ribary:1 neglect:1 build:5 already:1 occurs:4 said:1 exhibit:2 gradient:1 dp:1 separate:5 attentional:41 simulated:2 link:5 street:1 majority:1 sci:2 cochlear:2 argue:1 toward:3 ratio:1 classed:1 rise:2 perform:1 neuron:2 situation:2 segregate:1 incorporated:2 dc:2 carlyon:3 intensity:1 ordinate:1 bk:2 required:1 connection:10 acoustic:5 anstis:1 trans:1 beyond:2 suggested:1 below:3 pattern:1 perception:3 psychophys:1 wz:2 max:1 reciprocally:1 shifting:2 event:4 force:1 representing:2 wik:2 improve:1 temporally:4 deemed:1 psychol:3 extract:1 segregated:3 interesting:1 filtering:2 versus:1 age:10 degree:2 consistent:3 bank:1 interfering:1 squashing:1 cooke:1 excitatory:11 periodicity:1 summary:5 supported:1 placed:1 course:1 asynchronous:1 salience:1 normalised:2 fall:5 attaching:1 leaky:9 regard:1 curve:1 dip:1 calculated:2 instructed:1 made:5 coincide:1 reorientation:1 implicitly:1 neurobiological:1 global:7 active:5 conceptual:1 physiologically:1 additionally:2 stimulated:3 channel:40 expanding:1 contributes:1 investigated:3 complex:24 did:1 spread:2 arrow:1 arise:1 slow:4 aid:1 fails:1 comput:1 perceptual:8 weighting:2 synchronise:1 showing:1 decay:2 physiological:1 evidence:2 grouping:11 sequential:2 importance:1 ihigh:3 simply:1 likely:1 prevents:2 correlogram:9 contained:1 binding:1 corresponds:1 loud:2 determines:3 ma:1 presentation:1 bang:1 towards:2 oscillator:41 labelled:1 diminished:1 called:1 experimental:3 select:1 synchronised:10 modulated:2 signifying:1 heaviside:1 phenomenon:2 |
1,220 | 2,111 | Computing Time Lower Bounds for
Recurrent Sigmoidal Neural Networks
Michael Schmitt
Lehrstuhl Mathematik und Informatik, Fakultat fUr Mathematik
Ruhr-Universitat Bochum, D- 44780 Bochum, Germany
[email protected]
Abstract
Recurrent neural networks of analog units are computers for realvalued functions. We study the time complexity of real computation in general recurrent neural networks. These have sigmoidal,
linear, and product units of unlimited order as nodes and no restrictions on the weights. For networks operating in discrete time,
we exhibit a family of functions with arbitrarily high complexity,
and we derive almost tight bounds on the time required to compute
these functions. Thus, evidence is given of the computational limitations that time-bounded analog recurrent neural networks are
subject to.
1
Introduction
Analog recurrent neural networks are known to have computational capabilities that
exceed those of classical Turing machines (see, e.g., Siegelmann and Sontag, 1995;
Kilian and Siegelmann, 1996; Siegelmann, 1999). Very little, however, is known
about their limitations. Among the rare results in this direction, for instance,
is the one of Sima and Orponen (2001) showing that continuous-time Hopfield
networks may require exponential time before converging to a stable state. This
bound, however, is expressed in terms of the size of the network and, hence, does
not apply to fixed-size networks with a given number of nodes. Other bounds
on the computational power of analog recurrent networks have been established by
Maass and Orponen (1998) and Maass and Sontag (1999). They show that discretetime recurrent neural networks recognize only a subset of the regular languages in
the presence of noise. This model of computation in recurrent networks, however,
receives its inputs as sequences. Therefore, computing time is not an issue since
the network halts when the input sequence terminates. Analog recurrent neural
networks, however, can also be run as "real" computers that get as input a vector
of real numbers and, after computing for a while, yield a real output value. No
results are available thus far regarding the time complexity of analog recurrent
neural networks with given size.
We investigate here the time complexity of discrete-time recurrent neural networks
that compute functions over the reals. As network nodes we allow sigmoidal units,
linear units, and product units- that is, monomials where the exponents are ad-
justable weights (Durbin and Rumelhart, 1989) . We study the complexity of real
computation in the sense of Blum et aI. (1998). That means, we consider real numbers as entities that are represented exactly and processed without restricting their
precision. Moreover, we do not assume that the information content of the network
weights is bounded (as done, e.g., in the works of Balcazar et aI. , 1997; Gavalda and
Siegelmann, 1999). With such a general type of network, the question arises which
functions can be computed with a given number of nodes and a limited amount of
time. In the following , we exhibit a family of real-valued functions ft, l 2: 1, in one
variable that is computed by some fixed size network in time O(l). Our main result
is, then, showing that every recurrent neural network computing the functions ft
requires at least time nW /4). Thus, we obtain almost tight time bounds for real
computation in recurrent neural networks.
2
Analog Computation in Recurrent Neural Networks
We study a very comprehensive type of discrete-time recurrent neural network that
we call general recurrent neural network (see Figure 1). For every k, n E N there is
a recurrent neural architecture consisting of k computation nodes YI , . . . , Yk and n
input nodes Xl , ... , x n . The size of a network is defined to be the number ofits computation nodes. The computation nodes form a fully connected recurrent network.
Every computation node also receives connections from every input node. The input
nodes play the role of the input variables of the system. All connections are parameterized by real-valued adjustable weights. There are three types of computation
nodes: product units, sigmoidal units, and linear units. Assume that computation
node i has connections from computation nodes weighted by Wil, ... ,Wi k and from
input nodes weighted by ViI, .. . ,Vi n. Let YI (t) , . . . ,Yk (t) and Xl (t), ... ,X n (t) be the
values of the computation nodes and input nodes at time t, respectively. If node i
is a product unit, it computes at time t + 1 the value
(1)
that is, after weighting them exponentially, the incoming values are multiplied.
Sigmoidal and linear units have an additional parameter associated with them, the
threshold or bias ()i . A sigmoidal unit computes the value
where (J is the standard sigmoid (J( z ) = 1/ (1
simply outputs the weighted sum
+ e- Z ).
If node i is a linear unit, it
We allow the networks to be heterogeneous, that is, they may contain all three types
of computation nodes simultaneously. Thus, this model encompasses a wide class of
network types considered in research and applications. For instance, architectures
have been proposed that include a second layer of linear computation nodes which
have no recurrent connections to computation nodes but serve as output nodes (see,
e.g. , Koiran and Sontag, 1998; Haykin, 1999; Siegelmann, 1999). It is clear that in
the definition given here, the linear units can function as these output nodes if the
weights of the outgoing connections are set to O. Also very common is the use
of sigmoidal units with higher-order as computation nodes in recurrent networks
(see, e.g., Omlin and Giles, 1996; Gavalda and Siegelmann, 1999; Carrasco et aI.,
2000). Obviously, the model here includes these higher-order networks as a special
case since the computation of a higher-order sigmoidal unit can be simulated by
first computing the higher-order terms using product units and then passing their
.
I
I
sigmoidal, product, and linear units
computation
nodes
.
Yl
Yk
t
input nodes
Xl
Xn
I
Figure 1: A general recurrent neural network of size k. Any computation node may
serve as output node.
outputs to a sigmoidal unit. Product units , however, are even more powerful than
higher-order terms since they allow to perform division operations using negative
weights. Moreover, if a negative input value is weighted by a non-integer weight,
the output of a product unit may be a complex number. We shall ensure here that
all computations are real-valued. Since we are mainly interested in lower bounds,
however, these bounds obviously remain valid if the computations of the networks
are extended to the complex domain.
We now define what it means that a recurrent neural network N computes a function
f : ~n --+ llt Assume that N has n input nodes and let x E ~n. Given tE N,
we say that N computes f(x) in t steps if after initializing at time 0 the input
nodes with x and the computation nodes with some fixed values, and performing t
computation steps as defined in Equations (1) , (2) , and (3) , one of the computation
nodes yields the value f(x). We assume that the input nodes remain unchanged
during the computation. We further say that N computes f in time t if for every
x E ~n , network N computes f in at most t steps. Note that t may depend
on f but must be independent of the input vector. We emphasize that this is
a very general definition of analog computation in recurrent neural networks. In
particular, we do not specify any definite output node but allow the output to occur
at any node. Moreover, it is not even required that the network reaches a stable
state, as with attractor or Hopfield networks. It is sufficient that the output value
appears at some point of the trajectory the network performs. A similar view of
computation in recurrent networks is captured in a model proposed by Maass et al.
(2001). Clearly, the lower bounds remain valid for more restrictive definitions of
analog computation that require output nodes or stable states. Moreover, they
hold for architectures that have no input nodes but receive their inputs as initial
values of the computation nodes. Thus, the bounds serve as lower bounds also for
the transition times between real-valued states of discrete-time dynamical systems
comprising the networks considered here.
Our main tool of investigation is the Vapnik-Chervonenkis dimension of neural
networks. It is defined as follows (see also Anthony and Bartlett, 1999): A dichotomy
of a set S ~ ~n is a partition of S into two disjoint subsets (So , Sd satisfying
So U S1 = S. A class :F of functions mapping ~n to {O, I} is said to shatter S if
for every dichotomy (So , Sd of S there is some f E :F that satisfies f(So) ~ {O}
and f(S1) ~ {I}. The Vapnik-Chervonenkis (VC) dimension of :F is defined as
4"'+4",IL
'I
-1---Y-2----Y-5~1
S~
output
Y5
Y4
Figure 2: A recurrent neural network computing the functions fl in time 2l
+ 1.
the largest number m such that there is a set of m elements shattered by F. A
neural network given in terms of an architecture represents a class of functions
obtained by assigning real numbers to all its adjustable parameters, that is, weights
and thresholds or a subset thereof. The output of the network is assumed to be
thresholded at some fixed constant so that the output values are binary. The VC
dimension of a neural network is then defined as the VC dimension of the class of
functions computed by this network.
In deriving lower bounds in the next section, we make use of the following result
on networks with product and sigmoidal units that has been previously established
(Schmitt, 2002). We emphasize that the only constraint on the parameters of the
product units is that they yield real-valued, that is, not complex-valued, functions.
This means further that the statement holds for networks of arbitrary order, that is,
it does not impose any restrictions on the magnitude of the weights of the product
units.
Proposition 1. (Schmitt, 2002, Theorem 2) Suppose N is a feedforward neural
network consisting of sigmoidal, product, and linear units. Let k be its size and W
the number of adjustable weights. The VC dimension of N restricted to real-valued
functions is at most 4(Wk)2 + 20Wk log(36Wk).
3
Bounds on Computing Time
We establish bounds on the time required by recurrent neural networks for computing a family of functions fl : JR -+ JR, l 2:: 1, where l can be considered as a measure
of the complexity of fl. Specifically, fl is defined in terms of a dynamical system as
the lth iterate of the logistic map ?>(x) = 4x(1 - x), that is,
fl(X)
{
= 1,
?>(x)
l
?>(fl- l (x))
l > 2.
We observe that there is a single recurrent network capable of computing every fl
in time O(l).
Lemma 2. There is a general recurrent neural network that computes fl in time
2l + 1 for every l.
Proof. The network is shown in Figure 2. It consists of linear and second-order
units. All computation nodes are initialized with 0, except Yl, which starts with 1
and outputs 0 during all following steps. The purpose of Yl is to let the input x
output
Figure 3: Network Nt.
enter node Y2 at time 1 and keep it away at later times. Clearly, the value fl (x)
results at node Y5 after 2l + 1 steps.
D
The network used for computing fl requires only linear and second-order units. The
following result shows that the established upper bound is asymptotically almost
tight, with a gap only of order four . Moreover, the lower bound holds for networks
of unrestricted order and with sigmoidal units.
Theorem 3. Every general recurrent neural network of size k requires at least time
cl l / 4 j k to compute function fl' where c> 0 is some constant.
Proof. The idea is to construct higher-order networks Nt of small size that have
comparatively large VC dimension. Such a network will consist of linear and product
units and hypothetical units that compute functions fJ for certain values of j. We
shall derive a lower bound on the VC dimension of these networks. Assuming that
the hypothetical units can be replaced by time-bounded general recurrent networks,
we determine an upper bound on the VC dimension of the resulting networks in
terms of size and computing time using an idea from Koiran and Sontag (1998) and
Proposition 1. The comparison of the lower and upper VC dimension bounds will
give an estimate of the time required for computing k
Network Nt, shown in Figure 3, is a feedforward network composed of three networks
? r(1) , JVI
? r(2) , JVI
.r(3) . E ach networ k JVI
? r(/1) ,J.L = 1, 2, 3 , h as l ?lnput no d es Xl'
(/1) .. . , x I(/1)
JVI
and 2l + 2 computation nodes yb/1), ... , Y~r~l (see Figure 4). There is only one
adjustable parameter in Nt, denoted w, all other weights are fixed. The computation
nodes are defined as follows (omitting time parameter t):
for J.L
= 3,
for J.L = 1,2,
y~/1)
fll'--1 (Y~~)l) for i = 1, ... ,l and J.L = 1,2,3,
y}~{
y~/1) . x~/1), for i = 1, .. . ,l and
(/1)
Y21+l
(/1)
YIH
+ ... + Y21(/1)
J.L = 1,2,3,
c
- 1 2 3
lor
J.L , , ?
The nodes Yb/1) can be considered as additional input nodes for N//1), where N;(3)
gets this input from w, and N;(/1) from N;(/1+l) for J.L = 1,2. Node Y~r~l is the
output node of N;(/1), and node Y~~~l is also the output node of Nt. Thus, the entire
network has 3l + 6 nodes that are linear or product units and 3l nodes that compute
functions h, fl' or f12.
output
8
r - - - - - - - - - - - - '.....L - - - - - - - - - - - ,
I
I
B
B
t
t
I x~p)1
~
----t
input:
w
or
output of N;(P+1)
Figure 4: Network N;(p).
We show that Ni shatters some set of cardinality [3, in particular, the set S = ({ ei :
i = 1, . .. , [})3, where ei E {O, 1}1 is the unit vector with a 1 in position i and
elsewhere. Every dichotomy of S can be programmed into the network parameter
w using the following fact about the logistic function ? (see Koiran and Sontag,
1998, Lemma 2): For every binary vector b E {O, l}m, b = b1 .?. bm , there is some
real number w E [0,1] such that for i = 1, ... , m
?
E
{
[0,1 /2)
if bi = 0,
(1/2,1]
if bi = 1.
Hence, for every dichotomy (So, Sd of S the parameter w can be chosen such that
every (ei1' ei2 , ei3) E S satisfies
1/2 if (eillei2,eis) E So,
1/2 if (eillei2,eiJ E S1.
Since h +i2 H i 3 .12 (w) = ?i1 (?i2'1 (?i3 .1 2(w))), this is the value computed by Ni on
input (eill ei2' ei3), where ei" is the input given to network N;(p). (Input ei" selects
the function li"'I,,-1 in N;(p).) Hence, S is shattered by Ni, implying that Ni has
VC dimension at least [3.
Assume now that Ii can be computed by a general recurrent neural network of size
at most kj in time tj. Using an idea of Koiran and Sontag (1998), we unfold the
network to obtain a feedforward network of size at most kjtj computing fj. Thus we
can replace the nodes computing ft, ft, fl2 in Nz by networks of size k1t1, kltl, k12t12,
respectively, such that we have a feedforward network
consisting of sigmoidal,
product, and linear units. Since there are 3l units in Nl computing ft, ft, or fl2
and at most 3l + 6 product and linear units, the size of Nt is at most c1lkl2tl2
for some constant C1 > O. Using that Nt has one adjustable weight, we get from
Proposition 1 that its VC dimension is at most c2l2kr2tr2 for some constant C2 > o.
On the other hand, since Nz and Nt both shatter S, the VC dimension of Nt is at
least l3. Hence, l3 ~ C2l2 kr2 tr2 holds, which implies that tl2 2: cl 1/ 2/ kl2 for some
c > 0, and hence tl 2: cl 1/ 4/ kl .
D
'!J
Lemma 2 shows that a single recurrent network is capable of computing every
function fl in time O(l). The following consequence of Theorem 3 establishes that
this bound cannot be much improved.
Corollary 4. Every general recurrent neural network requires at least time 0(ll /4 )
to compute the functions fl.
4
Conclusions and Perspectives
We have established bounds on the computing time of analog recurrent neural
networks. The result shows that for every network of given size there are functions
of arbitrarily high time complexity. This fact does not rely on a bound on the
magnitude of weights. We have derived upper and lower bounds that are rather
tight- with a polynomial gap of order four- and hold for the computation of a
specific family of real-valued functions in one variable. Interestingly, the upper
bound is shown using second-order networks without sigmoidal units, whereas the
lower bound is valid even for networks with sigmoidal units and arbitrary product
units. This indicates that adding these units might decrease the computing time
only marginally. The derivation made use of an upper bound on the VC dimension
of higher-order sigmoidal networks. This bound is not known to be optimal. Any
future improvement will therefore lead to a better lower bound on the computing
time.
We have focussed on product and sigmoidal units as nonlinear computing elements.
However, the construction presented here is generic. Thus, it is possible to derive
similar results for radial basis function units, models of spiking neurons, and other
unit types that are known to yield networks with bounded VC dimension. The
questions whether such results can be obtained for continuous-time networks and for
networks operating in the domain of complex numbers, are challenging. A further
assumption made here is that the networks compute the functions exactly. By a
more detailed analysis and using the fact that the shattering of sets requires the
outputs only to lie below or above some threshold, similar results can be obtained
for networks that approximate the functions more or less closely and for networks
that are subject to noise.
Acknowledgment
The author gratefully acknowledges funding from the Deutsche Forschungsgemeinschaft (DFG). This work was also supported in part by the ESPRIT Working Group
in Neural and Computational Learning II, NeuroCOLT2, No. 27150.
References
Anthony, M. and Bartlett, P. L. (1999). Neural Network Learning: Theoretical
Foundations. Cambridge University Press, Cambridge.
Balcazar, J. , Gavalda, R., and Siegelmann, H. T. (1997). Computational power of
neural networks: A characterization in terms of Kolmogorov complexity. IEEE
Transcations on Information Theory, 43: 1175- 1183.
Blum, L., Cucker, F. , Shub, M. , and Smale, S. (1998) . Complexity and Real Computation. Springer-Verlag, New York.
Carrasco, R. C., Forcada, M. L., Valdes-Munoz, M. A. , and Neco, R. P. (2000).
Stable encoding of finite state machines in discrete-time recurrent neural nets
with sigmoid units. Neural Computation, 12:2129- 2174.
Durbin, R. and Rumelhart, D. (1989). Product units: A computationally powerful and biologically plausible extension to backpropagation networks. Neural
Computation, 1:133- 142.
Gavalda, R. and Siegelmann, H. T . (1999) . Discontinuities in recurrent neural
networks. Neural Computation, 11:715- 745.
Haykin, S. (1999). Neural Networks : A Comprehensive Foundation. Prentice Hall,
Upper Saddle River, NJ , second edition.
Kilian, J. and Siegelmann, H. T. (1996). The dynamic universality of sigmoidal
neural networks. Information and Computation, 128:48- 56.
Koiran, P. and Sontag, E . D. (1998). Vapnik-Chervonenkis dimension of recurrent
neural networks. Discrete Applied Mathematics, 86:63- 79.
Maass, W., NatschUiger, T., and Markram, H. (2001). Real-time computing without
stable states: A new framework for neural computation based on perturbations.
Preprint.
Maass, W. and Orponen, P. (1998). On the effect of analog noise in discrete-time
analog computations. Neural Computation, 10:1071- 1095.
Maass, W. and Sontag, E . D. (1999). Analog neural nets with Gaussian or other
common noise distributions cannot recognize arbitrary regular languages. Neural
Computation, 11:771- 782.
amlin, C. W. and Giles, C. L. (1996). Constructing deterministic finite-state automata in recurrent neural networks. Journal of the Association for Computing
Machinery, 43:937- 972.
Schmitt, M. (2002). On the complexity of computing and learning with multiplicative neural networks. Neural Computation, 14. In press.
Siegelmann, H. T . (1999). Neural Networks and Analog Computation: Beyond the
Turing Limit. Progress in Theoretical Computer Science. Birkhiiuser, Boston.
Siegelmann, H. T. and Sontag, E. D. (1995). On the computational power of neural
nets. Journal of Computer and System Sciences, 50:132- 150.
Sima, J. and Orponen, P. (2001). Exponential transients in continuous-time
symmetric Hopfield nets. In Dorffner, G., Bischof, H. , and Hornik, K. , editors , Proceedings of the International Conference on Artificial Neural Networks
ICANN 2001, volume 2130 of Lecture Notes in Computer Science, pages 806- 813,
Springer, Berlin.
| 2111 |@word polynomial:1 ruhr:2 yih:1 initial:1 chervonenkis:3 orponen:4 interestingly:1 nt:9 assigning:1 universality:1 must:1 partition:1 implying:1 haykin:2 characterization:1 valdes:1 node:53 sigmoidal:19 lor:1 shatter:2 c2:1 consists:1 little:1 cardinality:1 bounded:4 moreover:5 deutsche:1 what:1 nj:1 every:16 hypothetical:2 exactly:2 esprit:1 unit:44 before:1 sd:3 limit:1 consequence:1 encoding:1 might:1 nz:2 challenging:1 limited:1 programmed:1 bi:2 acknowledgment:1 definite:1 backpropagation:1 unfold:1 radial:1 regular:2 bochum:3 get:3 cannot:2 prentice:1 kr2:1 restriction:2 map:1 deterministic:1 automaton:1 deriving:1 construction:1 play:1 suppose:1 element:2 rumelhart:2 satisfying:1 carrasco:2 ft:6 role:1 preprint:1 initializing:1 eis:1 connected:1 kilian:2 decrease:1 yk:3 und:1 complexity:10 wil:1 dynamic:1 depend:1 tight:4 serve:3 networ:1 division:1 ei2:2 basis:1 hopfield:3 represented:1 kolmogorov:1 derivation:1 artificial:1 dichotomy:4 valued:8 plausible:1 say:2 obviously:2 sequence:2 net:4 product:19 lnput:1 ach:1 derive:3 recurrent:38 progress:1 implies:1 direction:1 closely:1 vc:13 transient:1 balcazar:2 require:2 investigation:1 proposition:3 extension:1 hold:5 considered:4 hall:1 nw:1 mapping:1 koiran:5 purpose:1 largest:1 establishes:1 tool:1 weighted:4 clearly:2 gaussian:1 i3:1 rather:1 corollary:1 derived:1 improvement:1 fur:1 indicates:1 mainly:1 sense:1 shattered:2 entire:1 selects:1 interested:1 germany:1 comprising:1 i1:1 issue:1 among:1 denoted:1 exponent:1 special:1 construct:1 shattering:1 represents:1 future:1 composed:1 simultaneously:1 recognize:2 comprehensive:2 dfg:1 replaced:1 consisting:3 attractor:1 investigate:1 nl:1 tj:1 capable:2 machinery:1 initialized:1 theoretical:2 instance:2 giles:2 subset:3 rare:1 monomials:1 universitat:1 international:1 river:1 yl:3 cucker:1 michael:1 tr2:1 lmi:1 li:1 de:1 wk:3 includes:1 ad:1 vi:1 later:1 view:1 multiplicative:1 start:1 capability:1 f12:1 il:1 ni:4 yield:4 informatik:1 marginally:1 ei3:2 trajectory:1 llt:1 reach:1 definition:3 kl2:1 thereof:1 associated:1 proof:2 appears:1 jvi:4 higher:7 specify:1 lehrstuhl:1 improved:1 yb:2 done:1 hand:1 receives:2 working:1 ei:4 nonlinear:1 logistic:2 omitting:1 effect:1 contain:1 y2:1 hence:5 symmetric:1 maass:6 i2:2 sima:2 ll:1 during:2 performs:1 fj:2 funding:1 sigmoid:2 common:2 spiking:1 exponentially:1 volume:1 analog:14 association:1 cambridge:2 munoz:1 ai:3 enter:1 mathematics:1 language:2 gratefully:1 l3:2 stable:5 operating:2 perspective:1 certain:1 verlag:1 binary:2 arbitrarily:2 yi:2 captured:1 additional:2 unrestricted:1 impose:1 determine:1 neco:1 ii:2 halt:1 converging:1 y21:2 heterogeneous:1 c1:1 receive:1 whereas:1 subject:2 call:1 integer:1 presence:1 exceed:1 feedforward:4 forschungsgemeinschaft:1 iterate:1 architecture:4 regarding:1 idea:3 whether:1 bartlett:2 sontag:9 dorffner:1 passing:1 york:1 clear:1 detailed:1 amount:1 processed:1 discretetime:1 disjoint:1 discrete:7 shall:2 group:1 four:2 amlin:1 threshold:3 blum:2 shatters:1 thresholded:1 eill:1 asymptotically:1 sum:1 run:1 turing:2 parameterized:1 powerful:2 almost:3 family:4 bound:27 layer:1 fl:14 fll:1 durbin:2 occur:1 constraint:1 unlimited:1 performing:1 fl2:2 jr:2 terminates:1 remain:3 wi:1 biologically:1 s1:3 restricted:1 computationally:1 equation:1 mathematik:2 previously:1 ofits:1 available:1 operation:1 multiplied:1 apply:1 observe:1 away:1 generic:1 include:1 ensure:1 siegelmann:11 restrictive:1 establish:1 classical:1 comparatively:1 unchanged:1 question:2 neurocolt2:1 said:1 exhibit:2 simulated:1 entity:1 berlin:1 ei1:1 y5:2 assuming:1 y4:1 statement:1 smale:1 negative:2 shub:1 adjustable:5 perform:1 upper:7 neuron:1 finite:2 extended:1 perturbation:1 arbitrary:3 required:4 kl:1 connection:5 bischof:1 fakultat:1 established:4 discontinuity:1 beyond:1 dynamical:2 below:1 encompasses:1 power:3 omlin:1 transcations:1 rely:1 realvalued:1 acknowledges:1 kj:1 fully:1 lecture:1 limitation:2 foundation:2 sufficient:1 schmitt:4 editor:1 elsewhere:1 supported:1 bias:1 allow:4 wide:1 focussed:1 markram:1 dimension:15 xn:1 valid:3 transition:1 computes:7 author:1 made:2 bm:1 far:1 approximate:1 emphasize:2 uni:1 keep:1 incoming:1 b1:1 assumed:1 continuous:3 hornik:1 complex:4 cl:3 anthony:2 domain:2 constructing:1 icann:1 main:2 noise:4 edition:1 tl:1 precision:1 position:1 exponential:2 xl:4 lie:1 weighting:1 theorem:3 specific:1 showing:2 evidence:1 consist:1 restricting:1 vapnik:3 adding:1 magnitude:2 te:1 gap:2 boston:1 vii:1 simply:1 eij:1 saddle:1 expressed:1 springer:2 satisfies:2 lth:1 replace:1 content:1 specifically:1 except:1 lemma:3 e:1 tl2:1 arises:1 outgoing:1 |
1,221 | 2,112 | Activity Driven Adaptive Stochastic
Resonance
Gregor Wenning and Klaus Oberrnayer
Department of Electrical Engineering and Computer Science
Technical University of Berlin
Franklinstr. 28/29 , 10587 Berlin
{grewe , oby}@cs.tu-berlin.de
Abstract
Cortical neurons might be considered as threshold elements integrating in parallel many excitatory and inhibitory inputs. Due to
the apparent variability of cortical spike trains this yields a strongly
fluctuating membrane potential, such that threshold crossings are
highly irregular. Here we study how a neuron could maximize its
sensitivity w.r.t. a relatively small subset of excitatory input. Weak
signals embedded in fluctuations is the natural realm of stochastic
resonance. The neuron's response is described in a hazard-function
approximation applied to an Ornstein-Uhlenbeck process. We analytically derive an optimality criterium and give a learning rule
for the adjustment of the membrane fluctuations, such that the
sensitivity is maximal exploiting stochastic resonance. We show
that adaptation depends only on quantities that could easily be
estimated locally (in space and time) by the neuron. The main
results are compared with simulations of a biophysically more realistic neuron model.
1
Introduction
Energetical considerations [1] and measurements [2] suggest , that sub-threshold
inputs, i.e. inputs which on their own are not capable of driving a neuron , play an
important role in information processing. This implies that measures must be taken,
such that the relevant information which is contained in the inputs is amplified in
order to be transmitted. One way to increase the sensitivity of a threshold device is
the addition of noise. This phenomenon is called stochastic resonance (see [3] for a
review) , and has already been investigated and experimentally demonstrated in the
context of neural systems (e.g. [3, 4]). The optimal noise level, however , depends
on the distribution of the input signals, hence neurons must adapt their internal
noise levels when the statistics of the input is changing. Here we derive and explore
an activity depend ent learning rule which is intuitive and which only depends on
quantities (input and output rates) which a neuron could - in principle - estimate.
The paper is structured as follows. In section 2 we describe the neuron model and we
introduce the m embrane potential dynamics in its hazard function approximation.
In section 3 we characterize stochastic resonance in this model system and we calculate the optimal noise level as a function of t he input and output rates. In section 4
we introduce an activity dependent learning rule for optimally adjusting the internal noise level, demonstrate its usefulness by applying it to t he Ornstein-Uhlenbeck
neuron and relate the phenomenon of stochastic resonance to its experimentally
accessible signature: the adaptation of the neuron 's transfer function . Section 5
contains a comparison to the results from a biophysically more realistic neuron
model. Section 6, finally, concludes with a brief discussion.
2
The abstract Neuron Model
Figure 1 a) shows the basic model setup. A leaky integrate-and-fire neuron receives
.",
a)
~;=".~
train with rate As
"0
>8 >
{5
-5"
'-<
O/l
b)
0.8
c
0.7
;::
0.6
.~
rateAo
8
0.9
"
0.5
E
~
0.4
"-'
0
0.2
.0
0.1
~
/'IWn
I
.S
2 N balanced Poisson
spike trains with rates
As
;::
0.
0.3
00
0.2
0.4
0.6
0.8
average membrane potential
Figure 1: a)The basic model setup. For explanation see text . b) A family of
Arrhenius type hazard functions for different noise levels. 1 corresponds to the
threshold and values below 1 are subthreshold .
e
a "signal" input , which we assume to be a Poisson distributed spike train with a rate
As. The rate As is low enough , so that the membrane potential V of the neuron
remains sub-threshold and no output spikes are generated . For the following we
assume that the information the input and output of the neuron convey is coded by
its input and output rates As and Ao only. Sensitivity is then increased by adding
2N balanced excitatory and inhibitory "noise" inputs (N inputs each) with rates
An and Poisson distributed spikes . Balanced inputs [5, 6] were chosen , because they
do not affect t he average membrane potential and allow to separate the effect of
decreasing the distance of the neuron's operating point to the threshold potential
from the effect of increasing the variance of the noise. Signal and noise inputs
are coupled to t he neuron via synaptic weights Ws and Wn for the signal and noise
inputs . The threshold of the neuron is denoted bye. Without loss of generality the
membrane time-constant, the neuron 's resting potential, and the neuron 's threshold
are set to one, zero , and one , respectively.
If the total rate 2N An of incoming spikes on t he "noise" channel is large and the
individual coupling constants Wn are small , the dynamics of the m embrane potential
can b e approximated by an Ornstein-Uhlenbeck process,
dV
=-V
dt
+ J.l
dt
+
(J"
dW,
(1)
where drift J.l and variance (J" are given by J.l = wsA s and (J"2 = w1A s + 2NwYvAN,
and where dW describes a Gaussian noise process with m ean zero and variance
one [8]. Spike initiation is included by inserting an absorbing boundary with reset.
Equation (1) can b e solved an alytically for special cases [8], but here we opt for
a more versatile approximation (cf. [7]). In this approximation, the probability of
crossing the threshold , which is proportional to the instantaneous output rate of
the neuron , is described by an effective transfer function. In [7] several transfer
functions were compared in their performance, from which we choose an Arrheniustype function ,
Ao(t) = c exp{ _ (e - ~(t))2},
(2)
cr
e-
where
x(t) is the distance in voltage between the noise free trajectory of the
membrane potential x(t) and the threshold x(t) is calculated from eq. (1) without
its diffusion term. Note that x(t) is a function of As, c is a constant. Figure 1 b)
shows a family of Arrhenius type transfer functions for different noise levels cr.
3
e,
Stochastic Resonance in an Ornstein- Uhlenbeck Neuron
Several measures can be used to quantify the impact of noise on the quality of signal
transmission through threshold devices . A natural choice is the mutual information
[9] between the distributions p( As) and p( Ao) of input and output rates, which we
will discuss in section 4, see also figure 3f. In order to keep the analysis and the
derivation of the learning rule simple , however, we first consider a scenario, in which
a neuron should distinguish between two sub-threshold input rates As and As + ~s.
Optimal distinguishability is achieved if the difference ~o of the corresponding
output rates is maximal, i.e. if
~o =
/(As
+ ~ s) - /(As)
(3)
= max ,
where / is the transfer function given by eq. (2). Obviously there is a close connection between these two measures , because increasing both of them leads to an
increase in the entropy of p( Ao) .
Fig. 2 shows plots of the difference
0.16
~o
of output rates vs. the level of noise, cr , for
0.4
~s
0.14
~s
0.35
AS= 7
0.12
:5
0
<:::]
:5
0
0.1
<:::]
0.08
0.06
0.04
0.02
0.05
00
50
0 2
100
[per cent]
50
100
[per cent]
Figure 2: ~ o vs. cr 2 for two different base rates As = 2 (left) and 7 (right) and 10
different values of ~ s = 0.01 , 0.02 , ... , 0.1. cr 2 is given in per cent of the maximum
cr 2 = 2N W;An. The arrows above t he x-axis indicate the position of the maximum
according to eq. (3), the arrowh eads below the x-axis indicate the optimal value
computed using eq. (5) (67% and 25%). Parameters were: N = la , An = 7,
Ws = 0.1 , and Wn E [0, 0.1].
different rates As and different values of ~ s . All curves show a clear maximum at a
particular noise level. The optimal noise level increases wit h decreasing t he input
rate As, but is roughly independent of the difference ~ s as long as ~ s is small.
Therefore, one optimal noise level holds even if a neuron has to distinguish several
sub-threshold input rates - as long as these rates are clustered around a given base
rate As.
The optimal noise level for constant As (stationary states) is given by the condition
d
d(j2 (f(A s + ~ s) - f(As)) = 0 ,
(4)
where f is given by eq. (2). Equation (4) can be evaluated in the limit of small
values of ~ s using a Taylor expansion up to the second order. We obtain
(j;pt
=
2(1 - ws As)2
(5)
if the main part of the variance of the membrane potential is a result of the balanced
.
'" 2N WNAN
2 , (f
2 -- - (1log(Ao/C)
- W, A, )2 , eq. (2) , eq. (5)
mput
, l."
e.f
1 (j 2 '"
c . eq. (1)) . S'mce (jopt
is equivalent to 1 + 2 log( Ao (A; ;0"2)) = O. This shows that the optimal noise level
depends either only on As or on Ao(As; (j2), both are quantities which are locally
available at the cell.
4
Adaptive Stochastic Resonance
We now consider the case , that a neuron needs to adapt its internal noise level
because the base input rate As changes. A simple learning rule which converges to
the optimal noise level is given by
~(j2
=
- f
(j2
log( - 2-) ,
(j opt
(6)
where the learning parameter f determines the time-scale of adaptation . Inserting
the corresponding expressions for the actual and the optimal variance we obtain a
learning rule for the weights W n ,
~wn =
-f
I og ( ( 2NAnw; )2 ) .
2 1 - ws As
(7)
Note, t hat equivalent learning rules (in the sense of eq. (6)) can be formulat ed for
the number N of the noise inputs and for their rates An as well. The r.h. s. of eqs .
(6) and (7) depend only on quantities which are locally available at the neuron.
Fig. 3ab shows the stochastic adaptation of the noise level, using eq.
randomly distributed As which are clustered around a base rate.
(7) , to
Fig. 3c-f shows an application ofthe learning rule, eq. (7) to an Ornstein-Uhlenbeck
neuron whose noise level needs to adapt to three different base input rates. T he
figure shows t he base input rate As (Fig. 3a). In fig. 3b the adaptation of Wn
according to eq. (7) is shown (solid line), for comparison t he Wn which maximizes
eq. (3) is also displayed (dash ed dotted lin e). Mutual information was calculated
between a distribution of randomly chosen input rates which are clustered around
the base rate As. The Wn that maximizes mutual Information between input and
output rates is displayed in fig. 3d (dashed lin e). Fig. 3e shows the ratio ~ o / ~ s
computed by using eq. (3) and the Wn calculated with eq. (8) (dashed dotted line)
and the same ratio for the quadratic approximation. Fig. 3f shows the mutual
information between the input and output rates as a function of the changing w n .
I[n~[ :~--/
0
0
2500
3000
1500
2000
2500
3000
0~
500 1000 1500
0
2000
1000
1500
~I
I,rJ
2000
500
?
0.1 5
w
110
b)
n 0.1
I1S
0
0.05
00
10
AS
o:i,ek '
500
1000
1500
2000
2500
3000
W
O'h
n 0.1
a)
500
I
1000
d)
..
?
5
00
As
500
1000
1500
2000
2500
3000
':1C)
0
0
I
500
1000
?
1500
ri"
I I
2500
3000
2000 2500 3000
?
time[updatesteps1
time [update steps]
Figure 3: a) Input rates As are evenly distributed around a base rate with width
0.5, in each time step one As is presented . b) Application of the learning rule eq.
(7) to t he rates shown in a). Adaptation of the noise level to t hree different input
base rates As. c ) The three base rates As. d) Wn as a function of time according
to eq. (7) (solid line) , the optimal Wn that maximizes eq. (3) (dashed dotted line)
and the optimal Wn that maximizes the m ut ual information between t he input and
output rates (dashed). T he opt imal values of Wn as the quadratic approximation,
eq. (5) yield are indicated by the black arrows. e ) The ratio b.. o / b.. s computed
from eq. (3) (dashed dotted line) and t he quadratic approximation (solid line) . f)
Mut ual information between input and output rates as a function of base rate and
changing synaptic coupling constant W n . For calculating the mutual information
the input rates were chosen randomly from the interval [As - 0.25 , As + 0.25] in each
time step . Parameters as in fig . 2.
T he fig ure shows , that the learning rule, eq. (7) in t he quadratic approximation
leads to values for () which are near-optimal, and that optimizing the difference of
output rates leads to results similar to t he optimization of the m ut ual information .
5
Conductance based Model Neuron
To check if and how t he results from the abstract model carryover to a biophysically
mode realistic one we explore a modified Hodgkin-Huxley point neuron with an
additional A-Current (a slow potassium current) as in [11] . T he dynamics of the
membrane potential V is described by t he following equation
C~~
- gL(V(t ) - EL) -
!iNam~ h(t)(V -
ENa)
- !iKn(t)4(V - EK) - !iAa ~ b(t)(V - EK)
+ l syn
+ la pp,
(8)
the parameters can be found in the appendix. All parameters are kept fixed through-
a
a)
II~
"}
'N
?::l
10
c
tS
80 ,---------------------------------,
b)
70
60
50
::I
~
o
~ 5
~
peak
conductances
8
40
~
.~
30
.5
20
<t>
U
00'----- 0 ,""
2 =:::...0-.4""""'-=--,L---~--,~
,2---"1.4
~
~
'i3
10
00
2
4
6
B
10
noiselevel in multiples of peak conductances
Figure 4: a) Transfer function for the conductance based model neuron with additional balanced input , a = 1, 2, 3, 4 b ) Demonstration of SR for the conductance
based model neuron. The plot shows the resonance for two different base currents
lapp = 0.7 and lapp = 0.2 and a E [0, 10].
~
-I
~
&
'0
-~
E
90 ,------ - - - - - - - - -- -- - - - - - - - - - - ---,
7
a)
E3 -
-B
~
~
5
--&.,
4-
b
:2
.j
1
70 -
~ 60 -
~
3
=
0)
gp ao -
50 40 -
1G
_~30 ~
20 -
-~
10 -
:;::
0 0~----------~~------~~~1
0.5
I drift [n~]
?0~----------0
~
.5
~--~------~
s"U-othresl""1<:>ld p<:>te:n.t:ial ( 8 =
1
)
Figure 5: a) Optimal noise-level as a function ofthe base current in the conductance
based model. b) Optimal noise-level as a function of the noise-free membrane
potential in the abstract model.
out all shown data. As balanced input we choose an excitatory Poisson spike train
with rate Ane = 1750 Hz and an inhibitory spike train with rate Ani = 750 Hz .
These spike trains are coupled to the neuron via synapses resulting in a synaptic
current as in [12]
ls yn = ge(V(t) - E e) + gi(V(t) - Ei)).
(9)
Every time a spike arrives at the synapse the conductance is increased by its peak
conductance ge i and decreases afterwards exponentially like exp{ - _t_
, } . The corT e, t
I
responding parameters are ge = a * 0.02 * gL , gi = a * 0.0615 * gL. The common
factor a is varied in the simulations and adjusts the height of the peak conductances,
gL is the leak conductance given above . Excitatory and inhibitory input are called
balanced if the impact of a spike-train at threshold is the same for excitation and
inhibition
TegeAne(Ee - B)
with Te i
I
= - TigiAni(Ei -
B)
= ~ J!fooo ge,i(t)dt . Note that the factor a does cancel in eq .
ge,t
(10)
(10).
Fig. 4a displays transfer funct ions in the conductance b ased setting with balanced
input. A family of functions with varying p eak conduct ances for the balanced input
is drawn .
~
100 r-----~----~----~----~----~-----.
d)
rJJ
~
~
50
?OL'~--~--~----~--~~--~--~
50
'-J
100
150
200
250
300
,-...,
i':k"-:-..:~. :
'"
0
50
1~-:o
50
100
150
200
250
300
100
150
200
250
300
Figure 6: Adaptive SR in the conductance based model. a) Currents drawn from
a uniform distribution of width 0.2 nA centered around base currents of 3, 8, 1 nA
respectively. b) Optimal noise-level in terms of a. Optimality refers to a semi-linear
fit to the data of fig. 5a. c) adapting the peak conductances using a in a learning
rule like eg. (8). d) Difference in spike count , for base currents I ? 0.1 nA and
using a as specified in c) .
For studying SR in t he conductance based framework , we apply the same paradigm
as in the abstract model. Given a certain average membrane potential, which is
adjusted via injecting a current I (in nA), we calculate the difference in the output
rate given a certain difference in the average membrane potential (mediated via the
injected current) I ? t:.I. A demonstration of stochastic resonance in the conductance based neuron can be seen in fig. 4b. In fig. 5a the optimal noise-level, in
terms of multiples a of the peak conductances , is plotted versus all currents that
yield a sub-threshold membrane voltage. For comparison we give the corresponding
relationship for the abstract model in fig. 5b .
Fig. 6 shows the performance of the conductance based model using a learning
rule like eg. (7). Since we do not have an analytically derived expression for (J opt
in the conductance based case, the relation (Jopt (I) , necessary for using eg. (7),
corresponds to a semi-linear fit to the (a opt , I) relation in fig. 5a.
6
Conclusion and future directions
In our contribution we have shown , that a simple and activity driven learning rule
can be given for the adaptation of the optimal noise level in a stochastic resonance
setting. Th e results from the abstract framework are compared with results from a
conductance based model neuron. A biological plausible m echanism for implem enting adaptive stochastic resonance in conductance based neurons is currently under
investigation.
Acknowledgments
Supported by: Wellcome Trust (061113/Z/00)
App e ndix: Parame t ers for the conductance based mode l n e uron
=
:':2
=
somatic conductances/ion-channel properties: em
1.0
,gL 0.05 ~ ,gNa
100 ~,gJ( = 40 ~,gA = 20 ~,EL = -65 mV,ENa = 55 mV, EJ(
- 80 mV, TA
=
= 20 ms ,
synaptic coupling : Ee = 0 mV, Ei = -80 mV, Te = 5 ms , Ti = 10 ms ,
spike initiation: dh = ~ dn = ngo - n <jQ = ~
mCO = <>m<>-ti3m'
dt
Th
O:m
= -O.l(V
55)/18),
h oo = <>h":i3h'
n co = <>n+i3 n '
O:h
O:n
'
dt
Tn'
dt
TA
'
+ 30)/(exp( -O .l(V + 30)) -
1), f3m = 4exp( -(V
+
= 0.07exp(-(V + 44)/20), f3h = l/( exp(-O. l(V + 14)) + 1) ,
= -O.Ol(V + 34)/(exp( -O.l(V + 34)) -1) , f3n = 0.125exp( -(V +
44)/80)
a oo = l/(exp( -(V + 50)/20) + 1) , boo = l/(exp((V
Th = in/(O:h + f3h) , Tn = in/(O:n + f3n), in = 0.1
+ 80)/6) + 1),
References
[1] S. B. Laughlin, RR de Ruyter van Steveninck and J.C. Anderson, The metabolic
cost of neural information, Nature Neuroscience, 1(1) , 1998, p.36-41
[2] J. S. Anderson, I. Lampl, D. C. Gillespie and D. Ferster, The Contribution of Noise to
Contrast Invariance of Orientation Tuning in Cat Visua l Cortex, Science, 290 , 2000,
p.1968-1972
[3] L. Gammaitoni, P. Hanggi, P. Jung and F. Marchesoni, Stochastic Resonance R eviews
of modern Physics, 70(1) , 1998, p.223-287
[4] D . F. Russel, L. A. Wilkens and F. Moss, Use of behavioral stochastic resonance by
paddle fish for feeding, Nature, 402 , 1999, p.291-294
[5] M. N . Shadlen, and W. T. Newsome, The Variable Discharge of Cortical Neurons:
Implications for Connec tivity, Computation, and Information Coding, The Journal of
Neuroscience, 18 (10), 1998, p.3870-3896
[6] M.V. Tsodyks and T. Sejnowski, Rapid state switching in balanced cortical network
models, Network: Computation in Neural Systems, 6 , 1995, p.I11-124
[7] H. E. Plesser and W. Gerstner, Noise in Integrate-and-Fire Neurons: From Stochastic
Input to Escape Rates, Neural Computation, 12 , 2000, p.367-384
[8] H. C. Tuckwell, Introduction to theoretical neurobiology: volume 2 nonlinear and
stochastic theories, Cambridge University Press, 1998
[9] T. M. Cover and J. A. Thomas, Elements of Information Theory, Wiley Series in
Telecommunications, 1991, 2nd edition
[10] A.R Bulsara, and A. Zador, Threshold detection of wide band signals: a noise-induced
maximum in the mutual information., PRE, 54(3), 1996, R2185-2188
[11] O. Shriki , D. Hansel and H. Sompolinsky, Modeling neuronal networks in cortex by rate
models using th e current-frequency respons e properties of cortical cells, Soc. Neurosci.
Abstr., 24 , p.143, 1998
[12] E. Salinas and T .J. Sejnowski, Impa ct of Co rrelated Synaptic Input on Output Firing
Rate and Variability in Simple Neurona l Mode ls J. Neurosci. 20 , 2000, p.6193-6209
| 2112 |@word nd:1 simulation:2 t_:1 solid:3 versatile:1 ld:1 contains:1 series:1 cort:1 current:12 wilkens:1 must:2 realistic:3 plot:2 update:1 v:2 stationary:1 device:2 ial:1 inam:1 height:1 dn:1 behavioral:1 introduce:2 rapid:1 roughly:1 ol:2 f3h:2 decreasing:2 actual:1 increasing:2 maximizes:4 every:1 ti:1 f3m:1 yn:1 engineering:1 limit:1 switching:1 ure:1 fluctuation:2 firing:1 might:1 black:1 co:2 steveninck:1 acknowledgment:1 adapting:1 pre:1 integrating:1 refers:1 suggest:1 close:1 ga:1 context:1 applying:1 equivalent:2 demonstrated:1 neurona:1 zador:1 l:2 wit:1 rule:13 adjusts:1 gammaitoni:1 dw:2 discharge:1 pt:1 play:1 element:2 crossing:2 approximated:1 role:1 electrical:1 solved:1 calculate:2 tsodyks:1 sompolinsky:1 decrease:1 balanced:10 leak:1 dynamic:3 signature:1 depend:2 funct:1 easily:1 lapp:2 cat:1 w1a:1 train:8 derivation:1 describe:1 effective:1 sejnowski:2 oby:1 klaus:1 salina:1 apparent:1 whose:1 plausible:1 ikn:1 statistic:1 gi:2 gp:1 obviously:1 rr:1 maximal:2 reset:1 adaptation:7 tu:1 relevant:1 inserting:2 j2:4 amplified:1 formulat:1 intuitive:1 ent:1 exploiting:1 potassium:1 abstr:1 transmission:1 converges:1 derive:2 coupling:3 oo:2 eq:23 soc:1 c:1 implies:1 indicate:2 quantify:1 bulsara:1 direction:1 stochastic:16 centered:1 respons:1 feeding:1 ao:8 clustered:3 investigation:1 opt:5 biological:1 adjusted:1 hold:1 around:5 considered:1 exp:10 shriki:1 driving:1 injecting:1 currently:1 hansel:1 gaussian:1 modified:1 i3:2 cr:6 ej:1 og:1 voltage:2 varying:1 derived:1 check:1 contrast:1 sense:1 dependent:1 el:2 plesser:1 w:4 relation:2 jq:1 orientation:1 denoted:1 resonance:14 special:1 mutual:6 cancel:1 future:1 carryover:1 escape:1 modern:1 randomly:3 individual:1 mut:1 fire:2 ab:1 conductance:22 detection:1 highly:1 arrives:1 implication:1 capable:1 necessary:1 conduct:1 taylor:1 plotted:1 theoretical:1 increased:2 modeling:1 cover:1 newsome:1 cost:1 subset:1 i1s:1 usefulness:1 uniform:1 paddle:1 characterize:1 optimally:1 peak:6 sensitivity:4 accessible:1 physic:1 na:4 choose:2 ek:3 potential:14 de:2 coding:1 mv:5 ornstein:5 depends:4 parallel:1 contribution:2 variance:5 yield:3 subthreshold:1 ofthe:2 weak:1 biophysically:3 trajectory:1 app:1 synapsis:1 synaptic:5 ed:2 pp:1 frequency:1 adjusting:1 realm:1 ut:2 syn:1 mco:1 ta:2 dt:6 response:1 synapse:1 evaluated:1 strongly:1 generality:1 anderson:2 receives:1 ei:3 trust:1 nonlinear:1 mode:3 quality:1 indicated:1 effect:2 fooo:1 analytically:2 hence:1 tuckwell:1 eg:3 embrane:2 width:2 excitation:1 m:3 demonstrate:1 tn:2 consideration:1 instantaneous:1 common:1 absorbing:1 exponentially:1 volume:1 he:21 resting:1 measurement:1 cambridge:1 ena:2 tuning:1 cortex:2 operating:1 inhibition:1 gj:1 base:15 own:1 optimizing:1 driven:2 scenario:1 certain:2 initiation:2 criterium:1 transmitted:1 seen:1 additional:2 maximize:1 paradigm:1 signal:7 semi:2 dashed:5 ii:1 multiple:2 rj:1 afterwards:1 technical:1 adapt:3 long:2 hazard:3 lin:2 coded:1 impact:2 basic:2 poisson:4 ane:1 uhlenbeck:5 achieved:1 cell:2 irregular:1 ion:2 addition:1 interval:1 sr:3 hz:2 induced:1 ngo:1 ee:2 near:1 enough:1 wn:12 boo:1 affect:1 fit:2 ased:1 implem:1 expression:2 e3:1 clear:1 locally:3 band:1 inhibitory:4 fish:1 dotted:4 estimated:1 neuroscience:2 per:3 threshold:17 drawn:2 changing:3 ani:1 diffusion:1 kept:1 injected:1 franklinstr:1 hodgkin:1 telecommunication:1 family:3 appendix:1 ct:1 dash:1 distinguish:2 display:1 quadratic:4 activity:4 huxley:1 ri:1 optimality:2 relatively:1 department:1 structured:1 according:3 membrane:13 describes:1 jopt:2 em:1 dv:1 taken:1 wellcome:1 equation:3 remains:1 wsa:1 discus:1 count:1 ge:5 studying:1 available:2 apply:1 fluctuating:1 hat:1 cent:3 hree:1 responding:1 lampl:1 cf:1 thomas:1 calculating:1 gregor:1 already:1 quantity:4 spike:14 distance:2 separate:1 berlin:3 evenly:1 parame:1 gna:1 relationship:1 ratio:3 demonstration:2 setup:2 relate:1 neuron:38 t:1 displayed:2 neurobiology:1 variability:2 arrhenius:2 varied:1 somatic:1 drift:2 specified:1 connection:1 distinguishability:1 below:2 max:1 explanation:1 gillespie:1 natural:2 ual:3 imal:1 brief:1 axis:2 grewe:1 concludes:1 mediated:1 coupled:2 moss:1 text:1 review:1 embedded:1 loss:1 proportional:1 versus:1 integrate:2 shadlen:1 principle:1 metabolic:1 excitatory:5 jung:1 gl:5 bye:1 free:2 supported:1 visua:1 allow:1 laughlin:1 wide:1 leaky:1 distributed:4 van:1 boundary:1 calculated:3 cortical:5 curve:1 adaptive:4 eak:1 keep:1 ances:1 incoming:1 channel:2 transfer:7 ruyter:1 nature:2 ean:1 expansion:1 investigated:1 gerstner:1 main:2 neurosci:2 arrow:2 noise:36 edition:1 convey:1 neuronal:1 fig:17 slow:1 wiley:1 sub:5 position:1 rjj:1 uron:1 er:1 adding:1 i11:1 te:3 mce:1 entropy:1 explore:2 adjustment:1 contained:1 corresponds:2 determines:1 dh:1 russel:1 i3h:1 ferster:1 experimentally:2 change:1 included:1 total:1 called:2 f3n:2 invariance:1 la:2 internal:3 mput:1 phenomenon:2 |
1,222 | 2,113 | On the Generalization Ability
of On-line Learning Algorithms
Nicol`o Cesa-Bianchi
DTI, University of Milan
via Bramante 65
26013 Crema, Italy
[email protected]
Alex Conconi
DTI, University of Milan
via Bramante 65
26013 Crema, Italy
[email protected]
Claudio Gentile
DSI, University of Milan
via Comelico 39
20135 Milano, Italy
[email protected]
Abstract
In this paper we show that on-line algorithms for classification and regression can be naturally used to obtain hypotheses with good datadependent tail bounds on their risk. Our results are proven without requiring complicated concentration-of-measure arguments and they hold
for arbitrary on-line learning algorithms. Furthermore, when applied to
concrete on-line algorithms, our results yield tail bounds that in many
cases are comparable or better than the best known bounds.
1 Introduction
One of the main contributions of the recent statistical theories for regression and classification problems [21, 19] is the derivation of functionals of certain empirical quantities
(such as the sample error or the sample margin) that provide uniform risk bounds for all the
hypotheses in a certain class. This approach has some known weak points. First, obtaining
tight uniform risk bounds in terms of meaningful empirical quantities is generally a difficult task. Second, searching for the hypothesis minimizing a given empirical functional
is often computationally expensive and, furthermore, the minimizing algorithm is seldom
incremental (if new data is added to the training set then the algorithm needs be run again
from scratch).
On-line learning algorithms, such as the Perceptron algorithm [17], the Winnow algorithm [14], and their many variants [16, 6, 13, 10, 2, 9], are general methods for solving
classification and regression problems that can be used in a fully incremental fashion. That
is, they need (in most cases) a short time to process each new training example and adjust
their current hypothesis. While the behavior of these algorithms is well understood in the
so-called mistake bound model [14], where no assumptions are made on the way the training sequence is generated, there are fewer results concerning how to use these algorithms
to obtain hypotheses with small statistical risk.
Littlestone [15] proposed a method for obtaining small risk hypotheses from a run of an
arbitrary on-line algorithm by using a cross validation set to test each one of the hypotheses
generated during the run. This method does not require any convergence property of the online algorithm and provides risk tail bounds that are sharper than those obtainable choosing,
for instance, the hypothesis in the run that survived the longest. Helmbold, Warmuth,
and others [11, 6, 8] showed that, without using any cross-validation sets, one can obtain
expected risk bounds (as opposed to the more informative tail bounds) for a hypothesis
randomly drawn among those generated during the run.
In this paper we prove, via refinements and extensions of the previous analyses, that online algorithms naturally lead to good data-dependent tail bounds without employing the
complicated concentration-of-measure machinery needed by other frameworks [19]. In
particular we show how to obtain, from an arbitrary on-line algorithm, hypotheses whose
risk is close to
with high probability (Theorems 2 and 3), where is the amount of
training data and is a data-dependent quantity measuring the cumulative loss of the online algorithm on the actual training data. When applied to concrete algorithms, the loss
bound translates into a function of meaningful data-dependent quantities. For classification problems, the mistake bound for the -norm Perceptron algorithm yields a tail risk
bound in terms of the empirical distribution of the margins ? see (4). For regression problems, the square loss bound for ridge regression yields a tail risk bound in terms of the
eigenvalues of the Gram matrix ? see (5).
2 Preliminaries and notation
#%$&(')'('*+ !
be arbitrary sets and
. An example is a pair
, where is an
Let
instance belonging to and
is the label associated with . Random variables will
be denoted in upper case and their realizations will be in lower case. We let be the
pair of random variables
, where
and take values in
and , respectively.
Throughout the paper, we assume that data are generated i.i.d. according to an unknown
probability distribution over . All probabilities and expectations will be understood with
to denote the vectorrespect to this underlying distribution. We use the short-hand
valued random variable
.
1
1 ,
:DC 1 E GFIH
:
"!
-
,./0
2(3(#,457698;:<#,=>?@BA
A hypothesis is any (measurable) mapping from instances
to predictions
, where is a given decision space. The risk of is defined by
,
where
is a nonnegative loss function. Unless otherwise specified, we will
assume that takes values in
for some known
. The on-line algorithms we
investigate are defined within a well-known mathematical model, which is a generalization
of a learning model introduced by Littlestone [14] and Angluin [1]. Let a training sequence
be fixed. In this learning model, an on-line
algorithm processes the examples in
one at a time in trials, generating a sequence of
hypotheses
. At the beginning of the -th trial, the algorithm receives the
and uses its current hypothesis
to compute a prediction
instance
for the label associated with . Then, the true value of the label
is disclosed and
, measuring how bad is the prediction
the algorithm suffers a loss
for the label . Before the next trial begins, the algorithm generates a new hypothesis
which may or may not be equal to
. We measure the algorithm?s performance on by
its cumulative loss
8 JKLA
,
J MNKOM7P
Q ! RS$T
U$V>(')'('W(
X#YZ 0 !
! !
Q!
,\[]?, $ )'(')'*+, !
^
4_
,/_`*$
,4_`*$T/_Ba 1
]_
/_
b_
:<#,/_`W$&/_#V
<_B
,S_`W$&/_c
_
Q!,_
,/_`*$
d
Q ! 5 ! :<g, _`W$ _ >
_ >'
_fe.$
In our analysis, we will write h and ij[<(')'('Wi when we want to stress the fact that the
!
cumulative loss and the hypotheses of the on-line algorithm are functions of the random
sample ! . In particular, throughout the paper i [ will denote the (deterministic) initial
hypothesis of an arbitrary on-line algorithm and, for each kml7^nl , io_ will be a random
variable denoting the ^ -th hypothesis of the on-line algorithm and such that the value of
ij_B5$&)')'('*+ ! does not change upon changes in the values of p_fq.$(')'('W+ ! .
Our goal is to relate the risk of the hypotheses produced by an on-line algorithm running
on an i.i.d. sequence
to the cumulative loss
of the algorithm on that sequence.
r!
hs#r!t
hsB !t
The cumulative loss
will be our key empirical (data-dependent) quantity. Via our
analysis we will obtain bounds of the form
k k l
2)3 gi [ )'(')'*i ! hsB n!t
where gi [<(')'('Wi is a specific function of the sequence of hypotheses i9[])')'('Wi
!
!
produced by the algorithm, and is a suitable positive constant. We will see that for specific on-line algorithms the ratio hs#r!t & can be further bounded in terms of meaningful
empirical quantities.
Our method centers on the following simple concentration lemma about bounded losses.
:
J l:Zl K . Let an arbitrary
i [ (')'('Wi ! when it is run
Lemma 1 Let be an arbitrary bounded loss satisfying
on-line algorithm output (not necessarily distinct) hypotheses
on . Then for any
we have
n!
JjM l k
k l '
k d ! 2)3i _`*$ h K
_fe.$
Proof. For each ^ Gk<(')')'* , set /_`*$ 2)3 ij_`W$) :Lij_`*$<9_#V/_# . We have
k d ! _`*$ k d ! 2)3 i _`*$ h '
_fe.$
_fe.$
Furthermore, KOl
_`*$ lNK , since : takes values in 8 J?K A . Also,
698 \_`*$ _`W$+AS72)3g,4_`W$) =6 8 : g,\_`*$<9_gV/_# _`W$?Ab+ J
where _`*$ denotes the -algebra generated by %$(')')'*> _`*$ . A direct application of the
Hoeffding-Azuma inequality [3] to the bounded random variables S[b)'(')'* `W$ proves the
!
lemma.
!
"
3 Concentration for convex losses
In this section we investigate the risk of the average hypothesis
d
, k ! , _`W$
_ e.$
$#&%('
,\[<?, $ (')'('*?, !
1
1
1
Theorem 2 Let be convex and : C F 8 J\K A be convex in the first argument. Let an
arbitrary on-line algorithm for : output (not necessarily distinct) hypotheses i [ )')'('Wi
!
M k the following holds
when the algorithm is run on r! . Then for any J M
2)3( i h K k l '
where
are the hypotheses generated by some on-line algorithm run on
training examples.1 The average hypothesis generates valid predictions whenever the decision space is convex.
1
Notice that the last hypothesis
+-,
)
*
is not used in this average.
$ !_ e $ L: g,4_:`W$&/>G'
$! !_ e $ )2 3 g, _`W$
!
: ,/V
l
2(3 ,4 l
Proof. Since is convex in the first argument, by Jensen?s inequality we have
Taking expectation with respect to
yields
. Using the last inequality along with Lemma 1 yields the thesis.
?@
"
Q !c &&
!_fe.$ (2 3 gi _`W$
This theorem, which can be viewed as the tail bound version of the expected bound in [11],
implies that the risk of the average hypothesis is close to
for ?most? samples
. On the other hand, note that it is unlikely that
concentrates around
, at least without taking strong assumptions on the underlying on-line algorithm.
Q!
6L8 h A
An application of Theorem 2 will be shown is Section 5. Here we just note that by applying
this theorem to the Weighted Majority algorithm [16], we can prove a version of [5, Theorem 4] for the absolute loss without resorting to sophisticated concentration inequalities
(details in the full paper).
4 Penalized risk estimation for general losses
:
If the loss function is nonconvex (such as the 0-1 loss) then the risk of the average hypothesis cannot be bounded in the way shown in the previous section. However, the risk
of the best hypothesis, among those generated by the on-line algorithm, cannot be higher
than the average risk of the same hypotheses. Hence, Lemma 1 immediately tells us that,
at
under no conditions on the loss function other than boundedness, for most samples
least one of the hypotheses generated has risk close to
. In this section we give
a technique (Lemma 4) that, using a penalized risk estimate, finds with high probability
such a hypothesis. The argument used is a refinement of Littlestone?s method [15]. Unlike
Littlestone?s, our technique does not require a cross validation set. Therefore we are able
to obtain bounds on the risk whose main term is
, where is the size of the whole
set of examples available to the learning algorithm (i.e., training set plus validation set in
Littlestone?s paper). Similar observations are made in [4], though the analysis there does
actually refer only to randomized hypotheses with 0-1 loss (namely, to absolute loss).
Q!
Q !t
Q !c &
,S_ by
_
^ =^t5
where ^ is the length of the suffix Q _fq.$ (')')' Q of the training sequence that the on-line
!
algorithm had not seen yet when ,/_ was generated, _ is the cumulative loss of ,4_ on that
suffix, and
k
k
/
5
'
Our algorithm chooses the hypothesis ,o ,
_ , where
^
[
< 3 _ `W $ _ ^ ^t '
!
For the sake of simplicity, we will restrict to losses : with range 8 J(kVA . However, it should
Let us define the penalized risk estimate of hypothesis
be clear that losses taking values in arbitrary bounded real interval can be handled using
techniques similar to those shown in Section 3. We prove the following result.
Theorem 3 Let an arbitrary on-line algorithm output (not necessarily distinct) hypotheses
when it is run on . Then, for any
, the hypothesis chosen using
the penalized risk estimate based on satisfies
i [<(')')'i !
!
2)3 i h
JjM l k
k k l '
*
i
The proof of this theorem is based on the two following technical lemmas.
Lemma 4 Let an arbitrary on-line algorithm output (not necessarily distinct) hypotheses
when it is run on . Then for any
the following holds:
i [<(')')'i !
!
J M M k
2)3 i [ _ *` $ 2)3i _ ^t
l '
!
Proof. Let
<3 [ _ `*$ 2(3(ij_B
^t
S' Let further i i , h
h , and set for brevity !
hZ_
_ ^
h
r '
For any fixed J we have
2)3 i= 2(3(i
d `W$
l ! _ =^t%l S2)3 ij_B 2(3(i < S' (1)
_fe [
Now, if
_ =^t%l (
_ lO2)3 ij_# =^t
holds then either
or
2)3i (
)
2(3 gi _ 2)3i %M
hold. Hence for any fixed ^ we can write
_ ^t%l )*2)3 i _ 2)3i )
l _ l 2(3(gi _ ^tS 2(3(i _ 2(3(gi <
2(3(i *2) 3ij_g 2)3i <
2)3(gij_# 2)3i )%M )S\2)3ij_#O2)3 i ( <
l _ l 2(3(gi _ ^t
2)3i ( )
2)3(gi _ 2)3i %M S\2)3i _ O2)3 i <9'
Probability (3) is zero if
. Hence, plugging (2) into (1) we can write
2(3 i O2)3(gi (
(
d! `*$
l
_ l 2(3 gi _B ^t
2(3(gi
_fe [
d `*$
l k ! _ O2)3(gi _ =^t
_fe [
l k k
or
(2)
(3)
where in the last two inequalities we applied Chernoff-Hoeffding bounds.
"
Lemma 5 Let an arbitrary on-line algorithm output (not necessarily distinct) hypotheses
when it is run on . Then for any
the following holds:
i [ (')')'i !
[ _ ! `*$ 2)3ij_B
!
h
^t
J M M k
k
k
k
l '
Proof. We have
d! `*$
k
[ _ ! `W$ 2)3g,4_g =^t
%l f_ e [ g2)3 g,4_#
d `*$
k ! 2(3 #, _
_fe [
d `*$
M k ! 2(3 #,\_#
_fe [
`*$
d
!
k
l 2(3 #,\_#
_fe [
where the last inequality follows from !_fe.$ k ^ l
h
[ _ ! `*$ 2(3(ij_B ^t
d `W$
l k ! 2)3 ij_B h
_fe [
l
by Lemma 1 (with K Gk ).
=^t
k
d! `W$ k
_ e [ ^t
d! `W$ k k
_ e [ =^
k k
. Therefore
k
k
k
)
*
k
"
Proof (of Theorem 3). The proof follows by combining Lemma 4 and Lemma 5, and by
"
overapproximating the square root terms therein.
5 Applications
For the sake of concreteness we now sketch two generalization bounds which can be obtained through a direct application of our techniques.
_`*$ H
_
N=H C
lGk$
,/_`W$T
_g5
S_`*$V
_B% k] k
*L
, _`*$
_ 7
_
S_ 4_`*$ <_
_
! " #
h
k
r!
2(3( i
i
k %$'& (r> ! ) Ok )k * Ok $'& + (r> ! k k (4)
The -norm Perceptron algorithm [10, 9] is a linear threshold algorithm which keeps in the
-th trial a weight vector
. On instance
, the
algorithm predicts by
sign
,
where
and . If the algorithm?s prediction is wrong (i.e., if
) then the algorithm
performs the weight update
. Notice that
yields the classical
Perceptron algorithm [17]. On the other hand,
gets an algorithm which
performs like a multiplicative algorithm, such as the Normalized Winnow algorithm [10].
Applying Theorem 3 to the bound on the number of mistakes for the -norm Perceptron
algorithm shown in [9], we immediately obtain that, with probability at least
with
respect to the draw of the training sample , the risk
of the penalized estimator
is at most
^
*
$ &
) Y
J
and for any ( such that ( `*$ lRk . The margin-based quantity
!
in [20] and accounts for
_ e $
J\)k <_(
_ ) is called soft margin
the distribution of margin values achieved by the examples in Q ! with respect to hyperplane
( . Traditional data-dependent bounds using uniform convergence methods (e.g., [19]) are
typically expressed in terms of the sample margin (^aC5 _ (
_ l $ ) & & , i.e., in terms of
the fraction of training points whose margin is at most ) . The ratio
+ (" Q ! occurring
(r Q !tL
for any
*
-
in (4) has a similar flavor, though the two ratios are, in general, incomparable.
$ &
We remark that bound (4) does not have the extra log factors appearing in the analyses
based on uniform convergence. Furthermore, it is significantly better than the bound in [20]
whenever
is constant, which typically occurs when the data sequence is not linearly
separable.
H
As a second application, we consider the ridge regression algorithm [12] for square loss.
Assume
and
. This algorithm computes at the beginning of the
-th trial the vector
which minimizes *
, where
, where is
. On instance the algorithm predicts with
the ?clipping? function
if
,
if
and
are thus bounded by . We can apply
if
. The losses
for ridge regression
(see [22, 2]) and
Theorem 2 to the bound on the cumulative
loss
obtain that, with probability at least
with respect to the draw of the training sample
, the risk
of the average hypothesis estimator is at most
8 p 0 A
_`*$
NJ
_
/, _`*$<
r
L%$
r
L
Xl
Nl
_ , _*` $
_
h
k
!
2)3 i
i
k ( hs+(r+ ! d ! _ _
_fe.$
!_ e $ $ _ (
for any ( H , where hs(r+ !c
^
_`*e.$ $ $
_gp E_`W$
_g
Nl
L
,
#
#
k
(5)
_ , ^ Gk<(')')'* . Then simple
!
j
T
_
_fe.$ tk _ /
#
_
f
_
.
e
$
! ! #
where the _ ?s are the eigenvalues of
. The nonzero
eigenvalues of
!
!
!
! are the
. Risk bounds in terms of
same as the nonzero eigenvalues of the Gram matrix
! ! we defer to the full paper a
the eigenvalues of the Gram matrix were also derived in [23];
_ , denotes the determinant
-dimensional identity matrix and
is the transpose of .2 Let us
of matrix
is the
denote by
the matrix whose columns are the data vectors
linear algebra shows that
!
*
comparison between these results and ours. Finally, our bound applies also to kernel ridge
regression [18] by replacing the eigenvalues of
with the eigenvalues of the kernel
,
Gram matrix
, where is the kernel being considered.
_
k l ^+ l
! !
References
[1] Angluin, D. Queries and concept learning, Machine Learning, 2(4), 319-342, 1988.
[2] Azoury, K., and Warmuth, M. K. Relative loss bounds for on-line density estimation
with the exponential family of distributions, Machine Learning, 43:211?246, 2001.
[3] K. Azuma. Weighted sum of certain dependend random variables. Tohoku Mathematical Journal, 68, 357?367, 1967.
2
Using a slightly different linear regression algorithm, Forster and Warmuth [7] have proven a
sharper bound on the expected relative loss. In particular, they have exhibited an algorithm computing
,#"
,
")(+*-,/.102354768 9
!
;:=<
hypothesis
such
>@?BA "#C
AF
EHGHI that in expectation (over ) the relative risk $&%'
is bounded by D
.
[4] A. Blum, A. Kalai, and J. Langford. Beating the hold-out: bounds for k-fold and
progressive cross-validation. In 12th COLT, pages 203?208, 1999.
[5] S. Boucheron, G. Lugosi, and P. Massart. A sharp concentration inequality with
applications. Random Structures and Algorithms, 16, 277?292, 2000.
[6] N. Cesa-Bianchi, Y. Freund, D. Haussler, D. P. Helmbold, R. E. Schapire, and M. K.
Warmuth. How to use expert advice. Journal of the ACM, 44(3), 427?485, 1997.
[7] J. Forster, and M. K. Warmuth. Relative expected instantaneous loss bounds. 13th
COLT, 90?99, 2000.
[8] Y. Freund and R. Schapire. Large margin classification using the perceptron algorithm. Machine Learning, 37(3), 277?296, 1999.
[9] C. Gentile The robustness of the -norm algorithms. Manuscript, 2001. An extended
abstract (co-authored with N. Littlestone) appeared in 12th COLT, 1?11, 1999.
[10] A. J. Grove, N. Littlestone, and D. Schuurmans. General convergence results for
linear discriminant updates, Machine Learning, 43(3), 173?210, 2001.
[11] D. Helmbold and M. K. Warmuth. On weak learning. JCSS, 50(3), 551?573, June
1995.
[12] A. Hoerl, and R. Kennard, Ridge regression: biased estimation for nonorthogonal
problems. Technometrics, 12, 55?67, 1970.
[13] J. Kivinen and M. K. Warmuth. Additive versus exponentiated gradient updates for
linear prediction. Information and Computation, 132(1), 1?64, 1997.
[14] N. Littlestone. Learning quickly when irrelevant attributes abound: A new linearthreshold algorithm. Machine Learning, 2, 285?318, 1988.
[15] N. Littlestone. From on-line to batch learning. In 2nd COLT, 269?284, 1989.
[16] N. Littlestone and M. K. Warmuth. The weighted majority algorithm. Information
and Computation, 108(2), 212?261, 1994.
[17] F. Rosenblatt. Principles of neurodynamics: Perceptrons and the theory of brain
mechanisms. Spartan Books, Washington, D.C., 1962.
[18] C. Saunders, A. Gammerman, and V. Vovk. Ridge Regression Learning Algorithm in
Dual Variables, In 15th ICML, 1998.
[19] J. Shawe-Taylor, P. Bartlett, R. Williamson, and M. Anthony, Structural Risk Minimization over Data-dependent Hierarchies. IEEE Trans. IT, 44, 1926?1940, 1998.
[20] J. Shawe-Taylor and N. Cristianini, On the generalization of soft margin algorithms,
2000. NeuroCOLT2 Tech. Rep. 2000-082, http://www.neurocolt.org.
[21] V.N. Vapnik, Statistical learning theory. J. Wiley and Sons, NY, 1998.
[22] V. Vovk, Competitive on-line linear regression. In NIPS*10, 1998. Also: Tech. Rep.
Department of Computer Science, Royal Holloway, University of London, CSD-TR97-13, 1997.
[23] R. C. Williamson, J. Shawe-Taylor, B. Sch?olkopf and A. J. Smola, Sample Based
Generalization Bounds, IEEE Trans. IT, to appear.
| 2113 |@word h:6 trial:5 determinant:1 version:2 norm:4 nd:1 gfih:1 boundedness:1 initial:1 denoting:1 ours:1 o2:4 current:2 yet:1 additive:1 informative:1 update:3 fewer:1 warmuth:8 beginning:2 short:2 provides:1 org:1 mathematical:2 along:1 direct:2 prove:3 expected:4 behavior:1 brain:1 actual:1 abound:1 begin:1 notation:1 underlying:2 bounded:8 minimizes:1 nj:1 dti:4 wrong:1 zl:1 appear:1 before:1 positive:1 understood:2 crema:2 mistake:3 lugosi:1 plus:1 therein:1 co:1 range:1 survived:1 empirical:6 significantly:1 get:1 cannot:2 close:3 ij_:5 risk:28 applying:2 www:1 measurable:1 deterministic:1 center:1 convex:5 simplicity:1 immediately:2 helmbold:3 estimator:2 haussler:1 searching:1 hierarchy:1 us:1 hypothesis:40 expensive:1 satisfying:1 predicts:2 jcss:1 cristianini:1 tight:1 solving:1 algebra:2 upon:1 derivation:1 distinct:5 london:1 query:1 spartan:1 tell:1 choosing:1 saunders:1 whose:4 valued:1 otherwise:1 ability:1 gi:12 dependend:1 online:3 sequence:8 eigenvalue:7 combining:1 realization:1 milan:3 olkopf:1 convergence:4 generating:1 incremental:2 tk:1 ij:1 strong:1 implies:1 concentrate:1 attribute:1 milano:1 require:2 generalization:5 preliminary:1 extension:1 hold:7 around:1 considered:1 mapping:1 nonorthogonal:1 estimation:3 hoerl:1 label:4 weighted:3 minimization:1 kalai:1 claudio:1 derived:1 june:1 longest:1 fq:1 tech:2 dependent:6 suffix:2 unlikely:1 typically:2 linearthreshold:1 classification:5 among:2 colt:4 denoted:1 dual:1 equal:1 washington:1 bramante:2 chernoff:1 progressive:1 icml:1 others:1 randomly:1 ab:1 technometrics:1 investigate:2 adjust:1 nl:3 grove:1 machinery:1 unless:1 taylor:3 littlestone:10 instance:6 column:1 soft:2 measuring:2 clipping:1 uniform:4 lgk:1 chooses:1 density:1 randomized:1 e_:1 quickly:1 concrete:2 again:1 thesis:1 cesa:3 opposed:1 hoeffding:2 book:1 expert:1 account:1 lnk:1 multiplicative:1 root:1 competitive:1 complicated:2 defer:1 contribution:1 square:3 yield:6 weak:2 produced:2 suffers:1 whenever:2 naturally:2 associated:2 proof:7 obtainable:1 sophisticated:1 actually:1 manuscript:1 ok:2 higher:1 though:2 furthermore:4 just:1 smola:1 langford:1 hand:3 receives:1 sketch:1 replacing:1 requiring:1 true:1 normalized:1 concept:1 hence:3 boucheron:1 nonzero:2 during:2 stress:1 ridge:6 performs:2 instantaneous:1 functional:1 tail:8 refer:1 seldom:1 resorting:1 shawe:3 had:1 recent:1 showed:1 italy:3 winnow:2 irrelevant:1 certain:3 nonconvex:1 inequality:7 rep:2 seen:1 gentile:3 lrk:1 full:2 technical:1 af:1 cross:4 concerning:1 plugging:1 prediction:6 variant:1 regression:12 expectation:3 kernel:3 achieved:1 want:1 interval:1 sch:1 extra:1 biased:1 unlike:1 exhibited:1 massart:1 structural:1 restrict:1 incomparable:1 translates:1 tohoku:1 handled:1 bartlett:1 remark:1 generally:1 clear:1 amount:1 authored:1 angluin:2 schapire:2 http:1 notice:2 sign:1 rosenblatt:1 gammerman:1 write:3 key:1 threshold:1 blum:1 drawn:1 kol:1 concreteness:1 fraction:1 sum:1 run:11 throughout:2 family:1 i9:1 draw:2 decision:2 comparable:1 bound:33 fold:1 nonnegative:1 alex:1 sake:2 generates:2 argument:4 separable:1 department:1 according:1 belonging:1 slightly:1 son:1 computationally:1 mechanism:1 needed:1 b_:1 available:1 apply:1 appearing:1 batch:1 robustness:1 denotes:2 running:1 prof:1 classical:1 added:1 quantity:7 occurs:1 neurocolt2:1 concentration:6 traditional:1 forster:2 gradient:1 neurocolt:1 majority:2 s_:3 discriminant:1 length:1 ratio:3 minimizing:2 difficult:1 fe:15 sharper:2 relate:1 gk:3 ba:2 unknown:1 bianchi:3 upper:1 observation:1 extended:1 dc:1 arbitrary:12 sharp:1 introduced:1 pair:2 namely:1 specified:1 comelico:1 nip:1 trans:2 able:1 beating:1 azuma:2 appeared:1 royal:1 suitable:1 kivinen:1 nicol:1 relative:4 freund:2 fully:1 dsi:2 loss:28 proven:2 versus:1 validation:5 principle:1 l8:1 penalized:5 last:4 transpose:1 exponentiated:1 perceptron:6 taking:3 absolute:2 gram:4 cumulative:7 valid:1 computes:1 made:2 refinement:2 employing:1 functionals:1 keep:1 neurodynamics:1 f_:1 obtaining:2 schuurmans:1 williamson:2 necessarily:5 anthony:1 main:2 linearly:1 azoury:1 whole:1 csd:1 kennard:1 advice:1 fashion:1 ny:1 wiley:1 exponential:1 xl:1 kla:1 theorem:11 bad:1 specific:2 jensen:1 disclosed:1 vapnik:1 occurring:1 margin:9 flavor:1 expressed:1 datadependent:1 conconi:2 g2:1 applies:1 satisfies:1 acm:1 goal:1 viewed:1 identity:1 change:2 unimi:3 hyperplane:1 vovk:2 lemma:12 called:2 meaningful:3 perceptrons:1 holloway:1 brevity:1 scratch:1 |
1,223 | 2,114 | Analog Soft-Pattern-Matching Classifier
using Floating-Gate MOS Technology
Toshihiko YAMASAKI and Tadashi SHIBATA*
Department of Electronic Engineering, School of Engineering
*Department of Frontier Informatics, School of Frontier Science
The University of Tokyo
7-3-1 Hongo, Bunkyo-ku, Tokyo, 113-8656, Japan
[email protected], [email protected]
Abstract
A flexible pattern-matching analog classifier is presented in conjunction with a robust image representation algorithm called Principal Axes Projection (PAP). In the circuit, the functional form of
matching is configurable in terms of the peak position, the peak height
and the sharpness of the similarity evaluation. The test chip was fabricated in a 0.6-?m CMOS technology and successfully applied to
hand-written pattern recognition and medical radiograph analysis using
PAP as a feature extraction pre-processing step for robust image coding.
The separation and classification of overlapping patterns is also experimentally demonstrated.
1
I ntr o du c ti o n
Pattern classification using template matching techniques is a powerful tool in implementing human-like intelligent systems. However, the processing is computationally very expensive, consuming a lot of CPU time when implemented as software running on general-purpose computers. Therefore, software approaches are not
practical for real-time applications. For systems working in mobile environment, in
particular, they are not realistic because the memory and computational resources
are severely limited. The development of analog VLSI chips having a fully parallel
template matching architecture [1,2] would be a promising solution in such applications because they offer an opportunity of low-power operation as well as very
compact implementation.
In order to build a real human-like intelligent system, however, not only the pattern
representation algorithm but also the matching hardware itself needs to be made
flexible and robust in carrying out the pattern matching task. First of all,
two-dimensional patterns need to be represented by feature vectors having substantially reduced dimensions, while at the same time preserving the human perception
of similarity among patterns in the vector space mapping. For this purpose, an image representation algorithm called Principal Axes Projection (PAP) has been de-
veloped [3] and its robust nature in pattern recognition has been demonstrated in the
applications to medical radiograph analysis [3] and hand-written digits recognition
[4]. However, the demonstration so far was only carried out by computer simulation.
Regarding the matching hardware, high-flexibility analog template matching circuits
have been developed for PAP vector representation. The circuits are flexible in a
sense that the matching criteria (the weight to elements, the strictness in matching)
are configurable. In Ref. [5], the fundamental characteristics of the building block
circuits were presented, and their application to simple hand-written digits was presented in Ref. [6]. The purpose of this paper is to demonstrate the robust nature of
the hardware matching system by experiments. The classification of simple
hand-written patterns and the cephalometric landmark identification in gray-scale
medical radiographs have been carried out and successful results are presented. In
addition, multiple overlapping patterns can be separated without utilizing a priori
knowledge, which is one of the most difficult problems at present in artificial intelligence.
2
I ma g e re pr es e n tati on by P AP
PAP is a feature extraction technique using the edge information. The input image
(64x64 pixels) is first subjected to pixel-by-pixel spatial filtering operations to detect edges in four directions: horizontal (HR); vertical (VR); +45 degrees (+45); and
?45 degrees (-45). Each detected edge is represented by a binary flag and four edge
maps are generated. The two-dimensional bit array in an edge map is reduced to a
one-dimensional array of numerals by projection. The horizontal edge flags are accumulated in the horizontal direction and projected onto vertical axis. The vertical,
+45-degree and ?45-degree edge flags are similarly projected onto horizontal,
-45-degree and +45-degree axes, respectively. Therefore the method is called ?Principal Axes Projection (PAP)? [3,4]. Then each projection data set is series connected
in the order of HR, +45, VR, -45 to form a feature vector. Neighboring four elements are averaged and merged to one element and a 64-dimensional vector is finally obtained. This vector representation very well preserves the human perception
of similarity in the vector space. In the experiments below, we have further reduced
the feature vector to 16 dimensions by merging each set of four neighboring elements into one, without any significant degradation in performance.
3
C i r cui t c o nf i g ura ti ons
A B C VGG
A B C VGG
IOUT
IOUT
1
1
2
2
4
4
1
VIN
13
VIN
RST
RST
Figure 1: Schematic of vector element matching circuit: (a) pyramid (gain reduction) type; (b) plateau (feedback) type. The capacitor area ratio is indicated
in the figure.
The basic functional form of the similarity evaluation is generated by the shortcut
current flowing in a CMOS inverter as in Refs. [7,8,9]. However, their circuits were
utilized to form radial basis functions and only the peak position was programmable.
In our circuits, not only the peak position but also the peak height and the sharpness
of the peak response shape are made configurable to realize flexible matching operations [5].
Two types of the element matching circuit are shown in Fig. 1. They evaluate the
similarity between two vector elements. The result of the evaluation is given as an
output current (IOUT ) from the pMOS current mirror. The peak position is temporarily memorized by auto-zeroing of the CMOS inverter. The common-gate transistor
with VGG stabilizes the voltage supply to the inverter. By controlling the gate bias
VGG, the peak height can be changed. This corresponds to multiplying a weight factor to the element. The sharpness of the functional form is taken as the strictness of
the similarity evaluation. In the pyramid type circuit (Fig. 1(a)), the sharpness is
controlled by the gain reduction in the input. In the plateau type (Fig. 1(b)), the
output voltage of the inverter is fed back to input nodes and the sharpness changes
in accordance with the amount of the feedback.
!"!
#$%#&"# #
#
#
'(')"'
'
'
'
16-dimension
15-vector
matching circuit
Decoder
4.5mm
Time-domain Winner-Take-All
Figure 2: Schematic of n-dimensional vector matching circuit utilizing the
pyramid type vector element circuits.
Figure 3: Photomicrograph of
soft-pattern-matching classi fier
circuit.
The total matching score between input and template vectors is obtained by taking
the wired sum of all I OUT ?s from the element matching circuits as shown in Fig. 2. A
multiplier circuit as utilized in Ref. [8] was eliminated because the radial basis function is not suitable for the template matching using PAP vectors. I SUM, the sum of
IOUT ?s, is then sunk through the nMOS with the VRAMP input. This forms a current
comparator circuit, which compares I SUM and the sink current in the nMOS with
VRAMP . The VOUT nodes are connected to a time-domain Winner-Take-All circuit [9].
A common ramp down voltage is applied to the VRAMP nodes of all vector matching
circuits. When V RAMP is ramped down from V DD to 0V, the vector matching circuit
yielding the maximum ISUM firstly upsets and its output voltage (VOUT ) shows a
0-to-1 transition. The time-domain WTA circuit senses the first upsetting signal and
memorizes the location in the open-loop OR-tree architecture [10]. In this manner,
the maximum-likelihood template vector is easily identified.
The circuits were designed and fabricated in a 0.6-?m double-poly triple-metal
CMOS technology. Fig. 3 shows the photomicrograph of a pattern classifier circuit
for 16-dimensional vectors. It contains 15 vector matching circuits. One element
matching circuit occupies the area of 150?m x 110?m. In the latest design, however,
the area is reduced to 54 ?m x 68 ?m in the same technology by layout optimization.
Further area reduction is anticipated by employing high-K dielectric films for capacitors since the capacitors occupy a large area. The full functioning of the chip
was experimentally confirmed [6]. In the following experiments, the simple vector
matching circuit in Fig. 2 was utilized to investigate the response from each template vector instead of just detecting the winner using the full chip.
E x per i me n tal r e s u l ts a nd di s c us si o n
Vector-element matching circuit
?
& & &'1- ( . 0
((
!
!
!
$#%
"
"
?
) ) )/- *,+.+ 0
+
"
?
4.1
!
!
?
4
!
"
Figure 4: Measured characteristics: (a) pyramid type; (b) plateau type. V GG was
varied from 3.0V to 4.5V, and control signals A~C from 000 to 111 for sharpness control.
Fig. 4 shows the measured characteristics of vector-element matching circuits in
both linear and log plots. The peak position was set at 1.05V by auto-zeroing. The
peak height was altered by V GG. Also, the operation mode was altered from the
above-threshold region to the sub-threshold region by V GG. In the plateau type circuit (Fig. 4(b)), I OUT becomes constant around the peak position and the flat region
widens in proportion to the amount of feedback. This is because the inverter operates so as to keep the floating gate potential constant in the high-gain region of the
inverter as in the case of virtual ground of an operational amplifier.
4.2
Matching of simple hand-w ritten patterns
Fig. 5 demonstrates the matching results for the simple input patterns. 16 templates
were stored in the matching circuit and several hand-written pattern vectors were
presented to the circuit as inputs. A slight difference in the matching score is observed between the pyramid type and the plateau type, but the answers are correct
for both types. Fig. 6 shows the effect of sharpness variation. As the sharpness gets
steeper, all the scores decrease. However, the score ratios between the winner and
loosers are increased, thus enhancing the winner discrimination margins. The
matching results with varying operational regimes of the circuit are given in Fig. 7.
The circuit functions properly even in the sub-threshold regime, demonstrating the
opportunity of extremely low power operation.
Presented
Patterns
Template Patterns
Template Patterns
Best
Matched
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
Template #
Best
Matched
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
Template #
Figure 5: Result of simple pattern matching: (a) pyramid type (left) where gain
reduction level was set with ABC=010; (b) plateau type (right) where feedback
ratio was set with ABC=101.
&%$
?
# "
!
& $ $
% $
$
,.-0/102434:<; 50=?64>@ 7A B , ; 8C, , , -9, /9, 19, 29, 3
&
$' +
$' *
$' )
$' (
$ ,D-0/14203454647E, 8, , , -9, /, 19, 29, 3
:; =>@ A B ; C
Figure 6: Effect of sharpness variation in the pyramid type with ABC=010.
Template #4
VGG=4.0V
VGG=3.5V
VGG=3.0V
VGG=2.5V
Input Pattern
Best Matched
1 2 3 4 5 6 7 8 9 10111213141516
Template#
Figure 7: Matching results as a function of V GG. Correct results are obtained in
the sub-threshold regime as well as in the above-threshold regime (the pyramid
type was utilized).
4.3
Application to gray-scale medical radiograph analysis
In Fig. 8, are presented the result of cephalometric landmark identification experiments, where the Sella (pituitary gland) pattern search was carried out using the
same matching circuit. Since the 64-dimension PAP representation is essential for
grayscale image recognition, the 64-dimension vector was divided into four
16-dimension vectors and the matching scores were measured separately and then
summed up by off-chip calculation. The correct position was successfully identified
both in the above-threshold (Fig. 8(b)) and the sub-threshold (Fig. 8(c)) regimes
using the 14 learned vectors as templates. In the previous work [3], successful
search was demonstrated by the computer simulation.
!
?"$#
!
?"$#
%'&)(+*-,/.* 01-2 3-4658792 3*+:
%'&)(+*-,/.* 01-2 3-4658792 3*+:
; <8=>@?BAC9DFE D
?G
; <8=>@?HDDFE I
?G
Figure 8: Matching results of Sella search using pyramid type with ABC=000:
(a) input image; (b) above-threshold regime; (c) sub-threshold regime.
4.4
Separation of overlapping patterns
Suppose an unknown pattern is presented to the matching circuit. The pattern might
consist of a single or multiple overlapping patterns. Let X represent the input vector
and W1st the winner (best matched) vector obtained by the matching circuit. Let the
first matching trial be expressed as follows:
? W1st
1st trial: X ????
matching
Then, the residue vector (X-W1st) is generated. The subtraction is perfomed in the
vector space. When an element in the residue vector becomes negative, the value is
set to 0. Such operation is easily implemented using the floating gate technique.
Here, the residue was obtained by off-line calculation. If the input pattern is single,
the residue vector is meaningless: only the leftover edge information remains in the
residue vector. If the input consists of overlapping patterns, the edge information of
other patterns remains. If the residue vector is very small, we can expect that the
input is single. But in many cases, the residue vector is not so small due to the distortion in hand-written patterns. Thus, it is almost impossible to judge which is the
case only from the magnitude of the residue vector. Therefore, we proceed to the
second trial to find the second winner:
2nd trial: X ? W1st
matching
????
? W2nd
With the same sequence, the second residue vector (X-W1st-W2nd), the third
(X-W1st -W2nd-W3rd) and so forth are generated by repeating the winner subtraction
after each trial. Then, new template vectors are generated such as W1st +W2nd,
W1st +W2nd+W3rd, and so forth. If the input vector is that of a single pattern, the
matching score is the highest at W1st and the scores are lower at W1st +W2nd and
W1st +W2nd+W3rd. On the other hand, if the input vector is that of two overlapping
patterns, the score is the highest at W1st+W2nd. This procedure can be terminated
automatically when the new template composed of n overlapping patterns yields
lower score than that of n-1 overlapping patterns. In this manner, we are able to
know how many patterns are overlapping and what patterns are overlapping without
a priori knowledge. An example of separating multiple overlapping patterns is illustrated in Fig. 9.
Template Patterns
Presented Patterns
1st try:
2nd try:
3rd try:
Final try:
Best Matched
?
?
?
?
+
+
+
+
Figure 9: Experimental result illustrating the algorithm for separating overlapping patterns. The solid black bars indicate the winner locations.
Template Patterns
Presented
Patterns
Best
Matched
A
#1
B C
#2
+
D E F
#3
+
+
+
(a)
A
+
B
+
C
+
+
(b)
D
+
E
+
+
F
+
+
+
Figure 10: Measured results demonstrating separation of multiple overlapping
patterns: (a) result of separation and classification (A~F are depicted in (b));
(b) newly created templates such as W1st+W2nd, W1st +W2nd+W3rd, and so on.
Several other examples are shown in Fig. 10. Pattern #1 is correctly classified as a
single rectangle by yielding the higher score for single template than that for
W1st +W2nd. Pattern #3 consists of three overlapping patterns, but is erroneously
recognized as four overlapping patterns. However, the result is not against human
perception. When we look at pattern #3, a triangle is visible in the pattern. This
mistake is quite similar to that made by humans.
5
C on cl us i o ns
A soft-pattern matching circuit has been demonstrated in conjunction with a robust
image representation algorithm called PAP. The circuit has been successfully applied to hand-written pattern recognition and medical radiograph analysis. The recognition of overlapping patterns similar to human perception has been also experimentally demonstrated.
Acknowledgments
Test circuits were fabricated in the VDEC program (The Univ. of Tokyo), in collaboration with Rohm Corp. and Toppan Printing Corp. The work is partially supported by the Ministry of Education, Science, Sports and Culture under the
Grant-in-Aid for Scientific Research (No. 11305024) and by JST in the program of
CREST.
References
[1] G.T. Tuttle, S. Fallahi, and A.A. Abidi. (1993) An 8b CMOS Vector A/D Converter. in
ISSCC Tech. Digest, vol. 36, pp. 38-39. IEEE Press.
[2] G. Cauwenberghs and V. Pedroni. (1995) A Charge-Based CMOS Parallel Analog Vector
Quantizer. In G. Tesauro, D. S. Touretzky and T.K. Leen (eds.), Advances in Neural Information Processing Systems 7, pp. 779-786. Cambridge, MA: MIT Press.
[3] M. Yagi, M. Adachi, and T. Shibata. (2000) A Hardware-Friendly Soft-Computing Algorithm for Image Recognition. X European Signal Processing Conf., Sept. 4-8, 2000
(EUSIPCO 2000), Vol. 2, pp. 729-732, Tampere, Finland.
[4] M. Adachi and T. Shibata. (2001) Image Representation Algorithm Featuring Human
Perception of Similarity for Hardware Recognition Systems. In Proc. of the Int. Conf. on
Artificial Intelligence (IC-AI'2001), Ed. by H. R. Arabnia, Vol. I, 229-234 (CSREA Press,
ISDBN: 1-892512-78-5), Las Vegas, Nevada, USA, June 25-28, 2001.
[5] T. Yamasaki and T. Shibata. (2001) An Analog Similarity Evaluation Circuit Featuring
Variable Functional Forms. In Proc. IEEE Int. Symp. Circuits Syst. (ISCAS 2001), Vol. 3,
pp. III-561-564, Sydney, Australia, May. 7-9, 2001.
[6] T. Yamasaki, K. Yamamoto and T. Shibata. (2001) Analog Pattern Classifier with Flexible Matching Circuitry Based on Principal-Axis-Projection Vector Representation. In Proc.
27th European Solid-State Circ. Conf. (ESSCIRC 2001), Ed. by F. Dielacher and H. Grunbacher, pp. 212-215 (Frontier Group), Villach, Austria, September 18-20, 2001.
[7] J. Anderson, J. C. Platt, and D. B. Kirk. (1993) An Analog VLSI Chip for Radial Basis
Functions. In S. J. Hanson, J. D. Cowan, and C. L. Giles Eds., Advances in Neural Information Processing Systems 5, pp. 765-772., San Maetro, CA; Morgan Kaufmann.
[8] L. Theogarajan and L. A. Akers. (1996) A Multi-Dimentional Analog Gaussian Radial
Basis Circuit. In Proc. IEEE Int. Symp. Circuits Syst. (ISCAS ?96), Vol. 3, pp. III-543 -546
Atlanta, GA, USA, May, 1996.
[9] L. Theogarajan and L. A. Akers. (1997) A scalable low voltage analog Gaussian radial
basis circuit. IEEE Trans. on Circuits and Systems II, Volume 44, No. 11, pp. 977 ?979,
1997.
[10] K. Ito, M. Ogawa and T. Shibata. (2001) A High-Performance Time-Domain Winner-Take-All Circuit Employing OR-Tree Architecture. In Proc. 2001 Int. Conf. on Solid
State Devices and Materials (SSDM2001), pp. 94-95, Tokyo, Japan, Sep. 26-28, 2001.
| 2114 |@word trial:5 illustrating:1 proportion:1 nd:3 open:1 simulation:2 solid:3 reduction:4 series:1 score:10 contains:1 current:5 si:1 written:7 realize:1 visible:1 realistic:1 shape:1 designed:1 plot:1 discrimination:1 intelligence:2 device:1 detecting:1 quantizer:1 node:3 location:2 firstly:1 height:4 supply:1 consists:2 ramped:1 isscc:1 symp:2 manner:2 veloped:1 multi:1 automatically:1 cpu:1 becomes:2 matched:6 circuit:45 what:1 substantially:1 developed:1 fabricated:3 nf:1 ti:2 charge:1 friendly:1 classifier:4 demonstrates:1 platt:1 control:2 medical:5 grant:1 engineering:2 accordance:1 mistake:1 severely:1 eusipco:1 ap:1 might:1 black:1 limited:1 averaged:1 practical:1 acknowledgment:1 block:1 digit:2 procedure:1 area:5 matching:46 projection:6 pre:1 radial:5 get:1 onto:2 ga:1 impossible:1 map:2 demonstrated:5 latest:1 layout:1 sharpness:9 utilizing:2 array:2 hd:1 x64:1 variation:2 controlling:1 suppose:1 element:14 recognition:8 expensive:1 utilized:4 observed:1 abidi:1 region:4 connected:2 decrease:1 highest:2 environment:1 carrying:1 basis:5 sink:1 triangle:1 easily:2 sep:1 chip:6 represented:2 separated:1 univ:1 artificial:2 detected:1 fallahi:1 pmos:1 dielectric:1 quite:1 film:1 distortion:1 ramp:2 circ:1 itself:1 final:1 sequence:1 transistor:1 nevada:1 neighboring:2 loop:1 flexibility:1 forth:2 ogawa:1 rst:2 double:1 wired:1 cmos:6 ac:2 measured:4 school:2 sydney:1 implemented:2 judge:1 indicate:1 direction:2 merged:1 tokyo:6 correct:3 human:8 occupies:1 australia:1 jst:1 memorized:1 implementing:1 numeral:1 virtual:1 education:1 material:1 cephalometric:2 frontier:3 mm:1 tati:1 around:1 ground:1 ic:1 mapping:1 mo:1 circuitry:1 stabilizes:1 finland:1 inverter:6 purpose:3 proc:5 leftover:1 successfully:3 tool:1 mit:1 gaussian:2 mobile:1 voltage:5 varying:1 sunk:1 conjunction:2 ax:4 june:1 properly:1 likelihood:1 tech:1 sense:1 detect:1 accumulated:1 vlsi:2 shibata:7 pixel:3 classification:4 flexible:5 among:1 priori:2 development:1 toshihiko:1 spatial:1 summed:1 extraction:2 having:2 eliminated:1 look:1 anticipated:1 intelligent:2 composed:1 preserve:1 floating:3 iscas:2 sella:2 amplifier:1 atlanta:1 investigate:1 evaluation:5 yielding:2 sens:1 edge:9 culture:1 tree:2 yamamoto:1 re:1 increased:1 soft:4 giles:1 tuttle:1 successful:2 configurable:3 stored:1 answer:1 st:2 peak:11 fundamental:1 off:2 informatics:1 rohm:1 conf:4 japan:2 syst:2 potential:1 de:1 coding:1 int:4 toppan:1 radiograph:5 try:4 lot:1 memorizes:1 steeper:1 cauwenberghs:1 parallel:2 vin:2 kaufmann:1 characteristic:3 yield:1 vout:2 identification:2 multiplying:1 confirmed:1 classified:1 plateau:6 yamasaki:4 touretzky:1 ed:4 akers:2 against:1 pp:9 di:1 gain:4 newly:1 austria:1 knowledge:2 back:1 higher:1 flowing:1 response:2 leen:1 anderson:1 just:1 hand:9 working:1 horizontal:4 overlapping:16 mode:1 gland:1 indicated:1 gray:2 scientific:1 usa:2 effect:2 building:1 multiplier:1 functioning:1 isum:1 illustrated:1 vdec:1 criterion:1 gg:4 demonstrate:1 image:9 vega:1 common:2 functional:4 winner:10 jp:2 volume:1 analog:10 slight:1 significant:1 cambridge:1 ai:1 rd:1 similarly:1 zeroing:2 similarity:8 tesauro:1 corp:2 binary:1 iout:4 preserving:1 ministry:1 morgan:1 subtraction:2 recognized:1 signal:3 ii:1 full:2 multiple:4 ntr:1 calculation:2 offer:1 divided:1 controlled:1 schematic:2 scalable:1 basic:1 enhancing:1 dimentional:1 represent:1 pyramid:9 addition:1 strictness:2 separately:1 residue:9 meaningless:1 ura:1 cowan:1 capacitor:3 ee:1 iii:2 architecture:3 identified:2 converter:1 regarding:1 vgg:8 bunkyo:1 proceed:1 programmable:1 amount:2 repeating:1 hardware:5 reduced:4 occupy:1 per:1 correctly:1 vol:5 group:1 upset:1 four:6 adachi:2 threshold:9 demonstrating:2 photomicrograph:2 rectangle:1 pap:9 sum:4 powerful:1 almost:1 electronic:1 separation:4 bit:1 esscirc:1 software:2 flat:1 tal:1 erroneously:1 nmos:2 extremely:1 yagi:1 department:2 cui:1 wta:1 pr:1 taken:1 computationally:1 resource:1 remains:2 know:1 subjected:1 fed:1 operation:6 gate:5 running:1 opportunity:2 widens:1 dfe:1 build:1 tadashi:1 digest:1 september:1 separating:2 landmark:2 decoder:1 me:1 ratio:3 demonstration:1 difficult:1 negative:1 ba:1 implementation:1 design:1 unknown:1 vertical:3 t:1 varied:1 hanson:1 learned:1 trans:1 able:1 bar:1 below:1 pattern:54 perception:5 regime:7 program:2 memory:1 power:2 suitable:1 hr:2 altered:2 technology:4 axis:2 created:1 carried:3 auto:2 sept:1 fully:1 expect:1 filtering:1 pedroni:1 triple:1 degree:6 metal:1 dd:1 collaboration:1 featuring:2 changed:1 supported:1 bias:1 template:21 taking:1 feedback:4 dimension:6 transition:1 made:3 projected:2 san:1 far:1 employing:2 crest:1 compact:1 hongo:1 ons:1 keep:1 consuming:1 grayscale:1 vramp:3 search:3 promising:1 ku:1 nature:2 robust:6 ca:1 operational:2 du:1 poly:1 cl:1 european:2 domain:4 terminated:1 ref:4 fig:16 vr:2 n:1 aid:1 sub:5 position:7 third:1 printing:1 kirk:1 ito:1 down:2 essential:1 consist:1 merging:1 mirror:1 magnitude:1 margin:1 depicted:1 tampere:1 expressed:1 temporarily:1 partially:1 sport:1 corresponds:1 abc:4 ma:2 comparator:1 shortcut:1 experimentally:3 change:1 operates:1 flag:3 principal:4 degradation:1 called:4 classi:1 total:1 e:1 experimental:1 la:1 evaluate:1 |
1,224 | 2,115 | 3 state neurons for contextual processing
Adam Kepecs* and Sridhar Raghavachari
Volen Center for Complex Systems
Brandeis University
Waltham MA 02454
{kepecs,sraghava}@brandeis.edu
Abstract
Neurons receive excitatory inputs via both fast AMPA and slow
NMDA type receptors. We find that neurons receiving input via
NMDA receptors can have two stable membrane states which are
input dependent. Action potentials can only be initiated from the
higher voltage state. Similar observations have been made in several brain areas which might be explained by our model. The interactions between the two kinds of inputs lead us to suggest that
some neurons may operate in 3 states: disabled, enabled and firing. Such enabled, but non-firing modes can be used to introduce
context-dependent processing in neural networks. We provide a
simple example and discuss possible implications for neuronal processing and response variability.
1
Introduction
Excitatory interactions between neurons are mediated by two classes of synapses:
AMPA and NMDA. AMPA synapses act on a fast time scale (TAMPA'" 5ms) , and
their role in shaping network dynamics has been extensively studied. The NMDA
type receptors are slow ((TNMDA '" 150ms) and have been mostly investigated for
their critical role in the induction of long term potentiation, which is thought to be
the mechanism for storing long term memories. Crucial to this is the unique voltage
dependence of NMDA receptors [6] that requires both the presynaptic neuron to
be active and the post-synaptic neuron to be depolarized for the channel to open.
However, pharamacological studies which block the NMDA receptors impair a variety of brain processes, suggesting that NMDA receptors also playa role in shaping
the dynamic activity of neural networks [10, 3, 8, 11, 2].
Therefore, we wanted to examine the role of NMDA receptors in post-synaptic
integration. Harsch and Robinson [4] have observed that injection of NMDA conductance that simulates synchronous synaptic input regularized firing while lowering response reliability. Our initial observations using a minimal model with
'The authors contributed equally to this work.
large NMDA inputs in a leaky dendrite showed a large regenerative depolarization. Neurons however, also possess a variety of potassium currents that are able
to limit these large excursions in voltage. In particular, recent observations show
that A-type potassium currents are abundant in dendrites of a variety of neurons
[7] . Combining these potassium currents with random NMDA inputs showed that
the membrane voltage alternated between two distinct subthreshold states. Similar
observations of two-state fluctuations have been made in vivo in several cortical
areas and the striatum [17, 9, 1]. The origin and possible functional relevance of
these fluctuations have remained a puzzle. We suggest that the NMDA type inputs
combined with potassium currents are sufficient to produce such membrane dynamics. Our results lead us to suggest that the fluctuations could be used to represent
contextual modulation of neuronal firing.
2
2 .1
NMDA-type input causes 2 state membrane fluctuations
Model
To examine the role of NMDA type inputs, we built a simple model of a cortical
neuron receiving AMPA and NMDA type inputs. To capture the spatial extent
of neuronal morphology we use a two-compartment model of pyramidal neurons
[15]. We represent the soma, proximal dendrites and the axon lumped into one
compartment containing the channels necessary for spike generation (INa and IK).
The dendritic compartment includes two potassium currents, a fast activating IKA
and the slower IKS along with a persistent sodium current INaP. The dendrite also
receives synaptic input as INMDA and IAMPA .
The membrane voltage of the neuron obeys the current balance equations:
while the dendritic voltage,
"\lid
obeys:
em
is the specific membrane capacitance which is taken to be 1 I1F / cm 2 for
where
both the dendrite and the soma for all cells and p =0.2, gc =0.05 determining the
electrotonic structure of the neuron.
The passive leak current in both the soma and dendrites were modeled as h eak =
El eak ), where gl eak was the leak conductance which was taken to be
0.3 mS/cm 2 for the soma and dendrite. El eak = -80mV was the leak reversal
potential for both the compartments. The voltage-dependent currents were modeled
according to the Hodgkin-Huxley formalism, with the gating variables obeying the
equation:
gl eak(V -
dx
(x oo(V) - x)
dt = ?x(ax(V)(1 - x) - ,sx(V)x) = ?x
Tx(V)
,
(3)
where x represents the activation/inactivation gates for the voltage-dependent currents.
The sodium current, INa = gNam~ h(VS - E Na ), where gNa = 45 mS/cm 2 and
sodium reversal potential, ENa = 55 mV with m oo(V) = a=(~)~~~(V). The
activation variables, O::m(V) = -O.l(V + 32)/[exp( -(V + 32)/10) - 1], 'sm(V) =
4exp( -[V + 57]/18); O::h(V) = 0.07 exp( -[V + 48]/20) and 'sh (V) = l/[exp( -{V +
18}/10) + 1], with ?m = ?h = 2.5.
The delayed rectifier potassium current, IKDr = gKn4(VS - EK), where gK = 9
mS/cm 2 and potassium reversal potential, EK = -80 mV with O::n(V) = -O.Ol(V +
34)/[exp( -(V + 34)/10) - 1], 'sn(V) = 0.125 exp( -[V + 44]/80), with ?n = 2.5.
In the dendrite, the persistent sodium current, INaP = gNapr~(V)(V - VNa ), with
roo(V) = 1/(1 + exp( -(V + 57)/5)) and gNaP =0.25 mS/cm 2 ? The two potassium
currents were hs = gKsq(V - VK), with qoo (V) = 1/(1 + exp( -(V + 50)/2)) and
Tq(V) = 200/(exp( -(V + 60)/10) + exp((V + 60)/10)) and gKS = 0.1 mS/cm 2 ; and
hA = gKAa~ (V)b(V - VK), with aoo(V) = 1/(1 + exp(-(V + 45)/6)), boo (V) =
1/(1 + exp(-(V + 56)/15)) and Tb(V) = 2.5(1 + exp((V + 60)/30)) and gKA = 10
mS/cm 2 .
The NMDA current, INMDA = fgNMDAS(V - ENMDA)/(l + 0.3[Mg] exp( -0.08V)),
where S was the activation variable and f denoted the inactivation of NMDA channels due to calcium entry. AMPA and NMDA inputs were modeled as conductance
kicks that decayed with TAMPA = 5 ms and TNMDA = 150 ms. Calcium dependent inactivation of the NMDA conductance was modeled as a negative feedback
df /dt = (foo - f)/2 , where f oo was a shallow sigmoid function that was 1 below a
conductance threshold of 2 ms/cm 2 and was inversely proportional to the NMDA
conductance above threshold. The coupling conductance is gc =0.1 mS/cm 2. The
asymmetry between the areas of the two compartments is taken into account in the
parameter p = somatic area/total area = 0.2. The temperature scaling factors
are ?h = ?n = 3.33. Other parameter values are: gLeak =0.3, gNa =36, gK =6 ,
gNaP =0.15, gKS =1, gKA =50 in mS/cm 2 unless otherwise noted; ELeak = -75,
ENa = +55, EK = -90, EKA = -80 in m V. Synchronous inputs were modeled as
a compound Poisson process representing 100 inputs firing at a rate A each spiking
with a probability of 0.1. Numerical integration was performed with a fourth-order
Runge-Kutta method using a 0.01 ms time step.
2.2
NMDA induced two-state fluctuations
Figure 1A shows the firing produced by inputs with high AMPA/NMDA ratio.
Figure 1B shows that the same spike train input delivered via synapses with a high
NMDA content results in robust two-state membrane behavior. We term the lower
and higher voltage states as UP and DOWN states respectively. Spikes caused by
AMPA-type inputs only occur during the up-state. In general, the same AMPA
input can only elicit spikes in the postsynaptic neuron when the NMDA input
switches that neuron into the up-state.
Transitions from down to up-state occur when synchronous NMDA inputs depolarize the membrane enough to cause the opening of additional NMDA receptor
channels (due to the voltage-dependence of their opening). This results in a regeneretive depolarization event, which is limited by the fast opening of IKA-type
Time [s]
Figure 1: Inputs with high AMPA-NMDA ratio cause the cell to spike (top trace,
=0.05, g N MDA =0.01). Strong NMDA inputs combined with potassium currents
(for the same AMPA input) result in fluctuations of the membrane potential between
two subthreshold states, with occasional firing due to the AMPA inputs (bottom trace,
gAMPA =0.01, gNMDA =0.1)
gA M PA
potassium channels. This up-state is stable because the regenerative nature and
long lifetime of NMDA receptor opening keeps the membrane depolarized, while
the slower I Ks potassium current prevents further depolarization. When input
ceases, NMDA channels eventually (TNMDA ~ 150ms) close and the membrane
jumps to the down-state. Note that while this bistable mechanism is intrinsic to
the membrane, it is also conditional upon input. Since the voltage threshold for
spike generation in the somal axon compartment is above the up-state, it acts as a
barrier. Thus, synchronous AMPA input in the down-state has a low probability of
eliciting a spike.
A number of previous experimental studies have reported similar phenomena in
various brain regions [16, 9, 1] where the two states persist even with all intrinsic
inward currents blocked but the inputs left intact [17] . Pharmacological block of
the potassium currents resulted in prolonged up-states [17] . These experimental
results suggested a conceptual model in which two-state fluctuations are (i) input
driven, (ii) the membrane states are stabilized by potassium currents. Nevertheless,
there remained a puzzle that (iii) up-state transitions are abrupt and (iv) the the
up-state is prolonged and restricted to a relatively narrow range of voltages. Our
model suggests a plausible mechanism for this phenomenon consistent with all experimental constraints. Below, we examine the origins of the two-state fluctuations
in light of these findings.
2.3
Analysis of two state fluctuations
Figure 2A shows the histogram of membrane potential for a neuron driven by combined AMPA and NMDA input at 30 Hz. There are two clear modes corresponding
to the up and down-states. The variability of the up-state and down-state voltages
is very low (u = 1.4 mV and 2.4 mV respectively) as observed. Figure 2B shows
the distribution of the up-state duration. The distribution of the up-state durations depend on the maximal NMDA conductance and the decay time constant
of NMDA (not shown), as well as the mean rate of NMDA inputs (Figure 2C).
A
B
C
O.
40
400
>-
~O.
30
U)
300
:c
E
'""
Q)
E
i=
CQ
.c
?0.
200
500
Time (ms)
100
1000
20 30 40 50
NMDA Rate (Hz)
Figure 2: A. Histogram of the up and down states. B. Dwell times of the up states C.
Mean duration of the up states increases with rate of NMDA inputs. Each histogram was
calculated over a run of 120 seconds.
Additionally, larger maximal potassium conductances shorten the duration of the
up states. Thus, we predict that the NMDA receptors are intimately involved in
shaping the firing characteristics of these neurons. Furthermore, our mechanistic
explanation leads a strong prediction about the functional role for these fluctuations
in neuronal processing.
3
Contextual processing with NMDA and AMPA pathways
Since NMDA and AMPA pathways have distinct roles in respectively switching and
firing our model neuron, we suggest the following conceptual model shown on Fig
3A. Without any input the neuron is at the rest or disabled state. Contextual input
(via NMDA receptors) can bring the neuron into an enabled state. Informational
(for instance, cue or positional) input (via AMPA receptors) can fire a neuron only
from this enabled state.
Where might such an architecture be used? In the CAl region of the hippocampus, pyramidal cells receive two distinct , spatially segregated input pathways: the
perforant path from cortex and the Schaffer collaterals from the CA3 region. The
perforant path has a very large NMDA receptor content [14] which is, interestingly, co-localized with high densities of I KA conductances [5]. Experimental [13]
and theoretical [12] observations suggest that these two pathways carry distinct
information. Lisman has suggested that the perforant path carries contextual information and the Schaffer collaterals bring sequence information [12]. Thus our model
seems to apply biophysically as well as suggest a possible way for CAl neurons to
carry out contextual computations. It is known that these cell can fire at specific
places in specific contexts. How might these different signals interact? As shown on
Fig3B , our model neuron can only fire spikes due to positional input when the right
context enables it. We note that a requirement for contextual processing is that the
two inputs be anatomically segregated, as they are in the CAl region. However,
we stress that the phenomenon of 2-state fluctuations itself is independent of the
location of the two kinds of inputs.
Figure 4A shows a similar processing scheme adapted for higher-order language
a. .;.~,. .,
A
Firing state
B'5
Context off
.~
<;;
~
~
0
Q.
""~ ~
0
~~'-)
?
~A/
'5
.~
Contextual input
Down-state / Disabled
Context on
<;;
~
~
0
Q.
g>
~~
Jll
Figure 3: A. Contextual input (high NMDA) switches the neuron from a rest state to
an up state. Informational input (high AMPA) cause the neuron to spike only from the
up state. B. Weak informational input can cause the cell to fire in conjunction with
the contextual input, (left traces) while strong informational input will not fire the cell
in the absence of contextual input (right traces). In this simulation, the soma/proximal
dendrite compartment receives AMPA input, while the NMDA input targets the dendritic
compartment.
processing. We simulated 3 neurons each receiving the same AMPA, informational
input. This might represent the word "green". Each of these neurons also receives
distinct contextual input via NMDA type receptors. These might, for instance,
represent specific noun groups: objects, people and fruit. The word "green" may
have very different meanings in these different contexts such as the color green, a
person who is a novice or an unripe fruit. We simulated this simple scenario shown
in Figure 4C. Even though each neuron receives the same strong AMPA input, their
firing seems uncorrelated. To evaluate the performance of the network in processing
contextual conjunctions, we measured the correlations between the information and
each contextual input. The most correlated at each moment was designated to be
the correct meaning. We then measured the number of spikes emitted by each
neuron during each "meaning" . Figure 4B shows that the neurons performed well ,
each tuned to fire preferentially in its appropriate context.
This simple example illustrates the use of a plausible biophysical mechanism for
computing conjuctions or multiplying with neurons.
4
Discussion
Voltage fluctuations between two subthresold levels with similar properties are observed in vivo in a variety of brain regions. Our model is in accordance with these
data and lead us to a new picture of how might these neuron operate in a functional
manner. Figure 3A shows our model operating as a 3-state device. It has a stable
low membrane state from which it cannot fire spikes, which we called disabled. It
also has a stable depolarized state from which action potentials can be elicited,
which this we call enabled state. Additionally, it has a firing state which is only
reachable from the enabled state.
What might be the role of the two non-firing states? We suggest that if high and
low NMDA-content pathways carry separate information these neurons can compute
A
"objects"
"people"
C
B
Contextual input:
90
0
"fruit"
111
0)(2)(3)
objects
people
frun
0
0
!6
o
J
~50
???
o 40
o
~
? 30
0
Sensory input:
0
"green"
o
2
4
0
1
2
3
I
1
2
3
1
Time(s)
Figure 4: A. Illustrative task for contextual processing in semantic inference. 3 neurons
each receive independent contextual (NMDA) and common informational (AMPA) input.
B. Voltage traces showing differences in firing patterns depending upon context. C. Each
neuron is tuned to its defined context. Correlation was measured between the informational
spike train and each contextual spike train smoothed with a gaussian filter (a = 60ms).
The most correlated context was defined to be the right one and the spikes of all neurons
were counted.
conjuctions, a simple form of multiplication. If the high NMDA-content pathway
carries contextual information then it would be in position to enable or disable a
neuron. In the enabled state, AMPA-type informational input could then fire a
neuron (Fig 3B).
We have presented a biophysical model for two-state fluctuations that is strongly
supported by data. One concern might be that most observations of 2-state fluctuations in vivo have been when the animal is anesthetized, implying that this kind
of neuronal dynamics is an artifact of the anesthetized state. However, these fluctuations have been observed in several different kinds of anesthesia, including local
anesthesia [16]. Furthermore, it has been shown that the duration of the up-states
correlate with orientation selectivity in visual cortical neurons suggesting that these
fluctuations might playa role in information processing. These observations suggest
that this phenomenon may be more indicative of a natural state of the cortex rather
than a by-product of anesthesia.
When the inputs with different AMPA/NMDA content are anatomically segregated,
t he NMDA input alone generates voltage fluctuations between a resting and depolarized state, while the AMPA input causes the neuron to spike when in the up-state.
This mechanism naturally leads to the suggestion that such two-state fluctuations
could have a function in computing context/input conjuctions. In summary, we
suggest the known biophysical mechanisms of some neurons can enable them two
operate as 3-state devices. In this mode of operation, the neurons could be used for
contextual processing.
Acknowledgments
We acknowledge John Lisman and John Fitzpatrick for useful discussion and suggestions.
2
3
References
[1] J. Anderson, 1. Lampl, 1. Reichova, M. Carandini, and D. Ferster. Stimulus dependence of two-state fluctuations of membrane potential in cat visual cortex. Nat
Neurosci, 3:617- 21 , 2000.
[2] A. Compte, N. BruneI, P. Goldman-Rakic, and X.J. Wang. Synaptic mechanisms and
network dynamics underlying spatial working memory in a cortical network model.
Cereb. Cortex, 10(9) :910- 923 , 2000 .
[3] S. Grillner, O. Ekeberg, A. Manira, A. Lansner, D. Parker, J. Tegner, and P. Wallen.
Intrinsic function of a neuronal network - a vertebrate central pattern generator.
Brain R es. Brain R es . R ev., 26:184- 197, 1998.
[4] A. Harsch and H.P.C. Robinson. Postsynaptic variability of firing rates in rat cortical
neurons: the role of input synchronization and synaptic nmda receptor conductance.
J. Neurosci., 20:6181- 6192, 2000.
[5] A Hoffman, JC Magee, CM Colbert, and D Johnston . K+ channel regulation of signal
propagation in dendrites of hippocampal pyramidal neurons. Nature, 387:869- 875,
1997.
[6] C.E. J ahr and C.F. Stevens. Voltage dependence of nmda-activated macroscopic
conductances predicted by single-channel kinetics. J Neurosci, 10:3178-82, 1990.
[7] D. Johnston, D.A. Hoffman, J .C. Magee, N.P. Poolos, S. Watanabe, C.M. Colbert,
and M. Migliore. Dendritic potassium channels in hippocampal pyramidal neurons.
J Physiol, 15:75- 81 , 2000.
[8] O. Kiehn and T. Eken. Functional role of plateau potentials in vertebrate motor
neurons. Curro Opin. Neurobiol., 8:746- 752 , 1998.
[9] B.L. Lewis and P . O 'Donnell . Ventral tegmental area afferents to the prefrontal
cortex maintain membrane potential 'up' states in pyramidal neurons via dl dopamine
receptors. Cereb. Cortex, 10:1168- 1175 , 2000.
[10] Y .X. Li, R. Bertram, and J . Rinzel. Modeling N-methyl-D-aspartate induced bursting
in dopamine neurons. N euroscience, 71(2):397- 410, 1996.
[11] J. Lisman , J.-M . Fellous, and X.J. Wang. A role for NMDA-receptor channels in
working memory. Nat. Neurosci. , 1(4):273- 275, 1998.
[12] J.E. Lisman. Relating hippocampal circuitry to function: recall of memory sequences
by reciprocal dentate-CA3 interactions. Neuron, 22:233- 242, 1999.
[13] B.L. McNaughton, C.A. Barnes, J. Meltzer, and R.J. Sutherland. Hippocampal granule cells are necessry for normal spatial learning but not for spatially-selective pyramidal cell discharge. Exp. Brain Res., 76:485- 496, 1989.
[14] N.A. Otmakhova and Lisman J. Dopamine selectively inhibits the direct cortical
pathway to the CAl hippocampal region. J Neurosci, 19:1437- 45 , 1999.
[1 5] P.F. Pinsky and J. Rinzel. Intrinsic and network rhythmogenesis in a reduced Traub
model for CA3 neurons. J. Comput. Neurosci. , 1:39- 60 , 1994.
[16] C.J. Wilson and P.M . Groves. Spontaneous firing patterns of identified spiny neurons
in the rat neostriatum. Brain Res, 220:67- 80, 1981.
[17] C.J. Wilson and Y. Kawaguchi. The origins of two-state spontaneuous fluctuations
of neostriatal spiny neurons. J. N eurosci., 16:2397- 2410, 1996.
| 2115 |@word h:1 hippocampus:1 seems:2 open:1 simulation:1 carry:5 moment:1 initial:1 tuned:2 interestingly:1 current:20 contextual:20 ka:1 activation:3 dx:1 john:2 physiol:1 numerical:1 aoo:1 enables:1 wanted:1 motor:1 opin:1 rinzel:2 v:2 implying:1 cue:1 alone:1 device:2 indicative:1 reciprocal:1 location:1 along:1 anesthesia:3 direct:1 ik:1 persistent:2 pathway:7 manner:1 introduce:1 behavior:1 examine:3 morphology:1 brain:8 ol:1 informational:8 goldman:1 prolonged:2 vertebrate:2 underlying:1 inward:1 what:1 kind:4 cm:11 neurobiol:1 depolarization:3 finding:1 act:2 sutherland:1 accordance:1 local:1 limit:1 striatum:1 receptor:17 switching:1 initiated:1 firing:16 fluctuation:20 modulation:1 path:3 might:9 studied:1 k:1 bursting:1 suggests:1 co:1 limited:1 range:1 obeys:2 unique:1 acknowledgment:1 block:2 area:6 elicit:1 thought:1 word:2 suggest:9 cannot:1 ga:1 close:1 cal:4 context:11 center:1 duration:5 abrupt:1 shorten:1 methyl:1 enabled:7 mcnaughton:1 discharge:1 target:1 spontaneous:1 origin:3 pa:1 persist:1 observed:4 role:12 bottom:1 wang:2 capture:1 region:6 lansner:1 leak:3 dynamic:5 depend:1 upon:2 various:1 tx:1 cat:1 train:3 distinct:5 fast:4 larger:1 plausible:2 roo:1 otherwise:1 itself:1 delivered:1 runge:1 sequence:2 mg:1 biophysical:3 interaction:3 maximal:2 product:1 combining:1 potassium:15 asymmetry:1 requirement:1 produce:1 adam:1 object:3 oo:3 coupling:1 depending:1 measured:3 strong:4 predicted:1 waltham:1 stevens:1 correct:1 filter:1 enable:2 bistable:1 gampa:1 potentiation:1 activating:1 dendritic:4 kinetics:1 normal:1 exp:15 gnmda:1 puzzle:2 predict:1 dentate:1 circuitry:1 fitzpatrick:1 ventral:1 eurosci:1 compte:1 hoffman:2 tegmental:1 gaussian:1 rather:1 inactivation:3 qoo:1 voltage:17 wilson:2 conjunction:2 ax:1 vk:2 inference:1 dependent:5 el:2 selective:1 orientation:1 denoted:1 animal:1 spatial:3 integration:2 noun:1 represents:1 stimulus:1 migliore:1 opening:4 resulted:1 delayed:1 fire:8 tq:1 maintain:1 conductance:12 brunei:1 kiehn:1 sh:1 traub:1 light:1 activated:1 implication:1 grove:1 necessary:1 collateral:2 rhythmogenesis:1 unless:1 iv:1 abundant:1 re:2 theoretical:1 minimal:1 instance:2 formalism:1 modeling:1 ca3:3 entry:1 reported:1 proximal:2 combined:3 person:1 decayed:1 density:1 donnell:1 off:1 receiving:3 gka:2 na:1 central:1 containing:1 prefrontal:1 wallen:1 ek:3 li:1 suggesting:2 potential:10 kepecs:2 account:1 includes:1 jc:1 caused:1 mv:5 afferent:1 performed:2 elicited:1 vivo:3 compartment:8 characteristic:1 who:1 subthreshold:2 weak:1 biophysically:1 produced:1 multiplying:1 plateau:1 synapsis:3 eka:1 synaptic:6 involved:1 naturally:1 carandini:1 gleak:1 recall:1 color:1 nmda:52 shaping:3 higher:3 dt:2 response:2 though:1 strongly:1 anderson:1 lifetime:1 furthermore:2 correlation:2 working:2 receives:4 propagation:1 jll:1 mode:3 artifact:1 disabled:4 depolarize:1 perforant:3 spatially:2 semantic:1 rakic:1 pharmacological:1 lumped:1 during:2 noted:1 illustrative:1 rat:2 m:17 hippocampal:5 stress:1 cereb:2 temperature:1 passive:1 bring:2 meaning:3 tegner:1 sigmoid:1 common:1 functional:4 spiking:1 he:1 resting:1 relating:1 blocked:1 ena:2 language:1 reliability:1 reachable:1 stable:4 cortex:6 operating:1 playa:2 showed:2 recent:1 driven:2 lisman:5 compound:1 scenario:1 selectivity:1 additional:1 disable:1 signal:2 ii:1 long:3 post:2 equally:1 regenerative:2 prediction:1 bertram:1 df:1 poisson:1 dopamine:3 histogram:3 represent:4 cell:8 receive:3 johnston:2 pyramidal:6 crucial:1 macroscopic:1 operate:3 depolarized:4 posse:1 rest:2 induced:2 hz:2 simulates:1 emitted:1 call:1 kick:1 iii:1 enough:1 boo:1 meltzer:1 variety:4 switch:2 architecture:1 identified:1 synchronous:4 cause:6 action:2 electrotonic:1 useful:1 clear:1 gks:2 extensively:1 reduced:1 stabilized:1 group:1 soma:5 threshold:3 nevertheless:1 lowering:1 run:1 fourth:1 hodgkin:1 place:1 excursion:1 curro:1 scaling:1 dwell:1 barnes:1 activity:1 mda:1 adapted:1 occur:2 constraint:1 huxley:1 generates:1 ekeberg:1 injection:1 relatively:1 inhibits:1 designated:1 according:1 membrane:17 em:1 postsynaptic:2 intimately:1 spiny:2 shallow:1 lid:1 explained:1 restricted:1 anatomically:2 taken:3 equation:2 discus:1 eventually:1 mechanism:7 mechanistic:1 reversal:3 operation:1 apply:1 occasional:1 appropriate:1 slower:2 gate:1 lampl:1 top:1 kawaguchi:1 granule:1 eliciting:1 iks:1 capacitance:1 spike:15 dependence:4 kutta:1 separate:1 simulated:2 presynaptic:1 extent:1 induction:1 gna:2 modeled:5 cq:1 ratio:2 balance:1 preferentially:1 regulation:1 mostly:1 gk:2 trace:5 negative:1 calcium:2 contributed:1 neuron:52 observation:7 tnmda:3 sm:1 acknowledge:1 neostriatum:1 colbert:2 variability:3 volen:1 gc:2 somatic:1 smoothed:1 schaffer:2 narrow:1 robinson:2 impair:1 able:1 suggested:2 below:2 pattern:3 ev:1 tb:1 tampa:2 built:1 memory:4 explanation:1 green:4 including:1 critical:1 event:1 natural:1 regularized:1 sodium:4 representing:1 scheme:1 inversely:1 picture:1 mediated:1 magee:2 alternated:1 sn:1 segregated:3 determining:1 multiplication:1 ina:2 synchronization:1 generation:2 suggestion:2 proportional:1 localized:1 generator:1 eleak:1 sufficient:1 consistent:1 fruit:3 storing:1 uncorrelated:1 excitatory:2 summary:1 gl:2 supported:1 barrier:1 anesthetized:2 leaky:1 feedback:1 calculated:1 cortical:6 transition:2 sensory:1 author:1 made:2 jump:1 novice:1 brandeis:2 counted:1 eak:5 correlate:1 keep:1 active:1 conceptual:2 vna:1 euroscience:1 additionally:2 channel:10 nature:2 robust:1 dendrite:10 interact:1 investigated:1 complex:1 ampa:24 neurosci:6 grillner:1 sridhar:1 ika:2 ahr:1 neuronal:6 fig:2 parker:1 slow:2 axon:2 foo:1 position:1 watanabe:1 obeying:1 comput:1 down:8 remained:2 fellous:1 specific:4 rectifier:1 gating:1 showing:1 aspartate:1 decay:1 cease:1 concern:1 dl:1 intrinsic:4 nat:2 illustrates:1 sx:1 positional:2 visual:2 prevents:1 lewis:1 ma:1 conditional:1 ferster:1 absence:1 content:5 total:1 called:1 experimental:4 e:2 intact:1 selectively:1 people:3 relevance:1 evaluate:1 phenomenon:4 correlated:2 |
1,225 | 2,116 | Batch Value Function Approximation via
Support Vectors
Thomas G Dietterich
Department of Computet Science
Oregon State University
Corvallis, OR, 97331
[email protected]
Xin W"ang
Department of Computer Science
Oregon State University
Corvallis, OR, 97331
wangxi@cs. orst. edu
Abstract
We present three ways of combining linear programming with the
kernel trick to find value function approximations for reinforcement
learning. One formulation is based on SVM regression; the second
is based on the Bellman equation; and the third seeks only to ensure
that good moves have an advantage over bad moves. All formulations attempt to minimize the number of support vectors while
fitting the data. Experiments in a difficult, synthetic maze problem
show that all three formulations give excellent performance, but the
advantage formulation is much easier to train. Unlike policy gradient methods, the kernel methods described here can easily 'adjust
the complexity of the function approximator to fit the complexity
of the value function.
1
Introduction
Virtually all existing work on value function approximation and policy-gradient
methods starts with a parameterized formula for the value function or policy and
then seeks to find the best policy that can be represented in that parameterized form.
This can give rise to very difficult search problems for which the Bellman equation
is of little or no use. In this paper, we take a different approach: rather than fixing
the form of the function approximator and searching for a representable policy, we
instead identify a good policy and then search for a function approximator that
can represent it. Our approach exploits the ability of mathematical programming
to represent a variety of constraints including those that derive from supervised
learning, from advantage learning (Baird, 1993), and from the Bellman equation. By
combining the kernel trick with mathematical programming, we obtain a function
approximator that seeks to find the smallest number of support vectors sufficient to
represent the desired policy. This side-steps the difficult problem of searching for
a good policy among those policies representable by a fixed function approximator.
Our method applies to any episodic MDP, but it works best in domains-such as
resource-constrained scheduling and other combinatorial optimization problemsthat are discrete and deterministic.
2
Preliminaries
There are two distinct reasons for studying value function approximation methods.
The primary reason is to be able to generalize from some set of training experiences
to produce a policy that can be applied in new states that were not visited during
training. For example, in Tesauro's (1995) work on backgammon, even after training
on 200,000 games, the TD-gammon system needed to be able to generalize to new
board positions that it had not previously visited. Similarly, in Zhang's (1995) work
on space shuttle scheduling, each individual scheduling problem visits only a finite
number of states, but the goal is to learn from a series of "training" problems and
generalize to new states that arise in "test" problems. Similar MDPs have been
studied by Moll, Barto, Perkins & Sutton (1999).
The second reason to study function approximation is to support learning in continuous state spaces. Consider a robot with sensors that return continuous values.
Even during training, it is unlikely that the same vector of sensor readings will ever
be experienced more thaIl once. Hence, generalization is critical during the learning
process as well as after learning.
The methods described in this paper address only the first of these reason. Specifically, we study the problem of generalizing from a partial policy to construct a
complete policy for a Markov Decision Problem (MDP). Formally, consider a discrete time MDP M with probability transition function P(s/ls, a) (probability that
state Sl will result from executing action a in state s) and expected reward function
R(s/ls, a) (expected reward received from executing action a in state s and entering
state Sl). We will assume that, as in backgammon and space shuttle scheduling,
P(s/ls, a) and R(s/ls, a) are known and available to the agent, but that the state
space is so large that it prevents methods such as value iteration or policy iteration from being applied. Let L be a set of "training" states for which we have an
approximation V(s) to the optimal value function V*(s), s E L. In some cases, we
will also assume the availability of a policy 'ff consistent with V(s). The goal is to
construct a parameterized approximation yes; 8) that can be applied to all states
in M to yield a good policy if via one-step lookahead search. In the experiments
reported below, the set L contains states that lie along trajectories from a small set
of "training" starting states So to terminal states. A successful learning method will
be able to generalize to give a good policy for new starting states not in So. This
was the situation that arose in space shuttle scheduling, where the set L contained
states that were visited while solving "training" problems and the learned value
function was applied to solve "test" problems.
To represent states for function approximation, let X (s) denote a vector of features
describing the state s. Let K(X 1 ,X2 ) be a kernel function (generalized inner product) of the two feature vectors Xl and X 2 ? In our experiments, we have employed
the gaussian kernel: K(X1,X2;U) == exp(-IIX1 - X 2 11 2 / ( 2 ) with parameter u.
3
Three LP Formulations of Function Approximation
We now introduce three linear programming formulations of the function approximationproblem. We first express each of these formulations in terms of a generic
fitted function approximator V. Then, we implement V(s) as the dot product of
a weight vector W with the feature vector X (s): V(s) == W . X (s). Finally, we
apply the "kernel trick" by first rewriting W as a weighted sum of the training
points Sj E L, W == ~j ajX(sj), (aj 2: 0), and then replacing all dot products
between data points by invocations of the kernel function K. We assume L con-
tains all states along the best paths from So to terminal states and also all states
that can be reached from these paths in one step and that have been visited during
exploration (so that V is known). In all three formulations we have employed linear
objective functions, but quadratic objectives like those employed in standard support vector machines could be used instead. All slack variables in these formulations
are constrained- to be non-negative.
Formulation 1: Supervised Learning. The first formulation treats the value
function approximation problem as a supervised learning problem and applies the
standard c-insensitive loss function (Vapnik, 2000) to fit the function approximator.
minimize
L
[u(s)
+ v(s)]
S
subject to V(s)
+ u(s) 2:: V(s) - c; V(s) - v(s) :::; V(s) + c
"Is E L
In this formulation, u(s) and v(s) are slack variables that are non-zero only if V(s)
has an absolute deviation from V(s) of more than c. The objective function seeks to
minimize these absolute deviation errors. A key idea of support vector methods is
to combine this objective function with a penalty on the norm of the weight vector.
We can write this as
minimize
IIWlll + C L[u(s) + v(s)]
S
subject to W? X(s)
+ u(s) 2:: V(s) -
c;
W? X(s) - v(s) :::; V(s)
+c
"Is E L
The parameter C expresses the tradeoff between fitting the data (by driving the
slack variables to zero) and minimizing the norm of the weight vector. We have
chosen to minimize the I-norm of the weight vector (11Wlll == Ei IWi!), because this
is easy to implement via linear programming. Of course, if the squared Euclidean
norm of W is preferred, then quadratic programming methods could be applied to
minimize this.
Next, we introduce the assumption that W can be written as a weighted sum of the
data points themselves. Substituting this into the constraint equations, we obtain
minimize
L
aj
+ C L[u(s) + v(s)]
j
subject to
Ej
Ej
8
ajX(sj) . X(s) + u(s) ~ V(s) - c
ajX(sj) . X(s) - v(s) :::; V(s) + c
"Is E L
"Is E L
Finally, we can apply the kernel trick by replacing each dot product by a call to a
kernel function:
minimize
Laj
+ CL[u(s) + v(s)]
j
subject to
Ej
Ej
s
ajK(X(sj),X(s)) + u(s) 2:: V(s) - c
ajK(X(sj), X(s)) - v(s) :::; V(s) + c
"Is E L
"Is E L
Formulation 2: Bellman Learning. The second formulation introduces constraints from the Bellman equation V(s) == maxa ESI P(s'ls, a)[R(s'ls, a) + V(s')].
The standard approach to solving MDPs via linear programming is the following.
For each state s and action a,
minimize
L u(s; a)
s,a
subject to V(s) == u(s,a)
+ LP(s'ls,a)[R(s'ls,a) + V(s')]
s'
The idea is' that for the optimal action a* in state s, the slack variable u(s, a*)
can be driven to zero, while for non-optimal actions a_, the slack u(s, a_) will
remain non-zero. Hence, the minimization of the slack variables implements the
maximization operation of the Bellman equation.
We attempted to apply this formulation with function approximation, but the errors
introduced by the approximation make the linear program infeasible, because V(s )
must sometimes be less than the backed-up value Ls' P(s'ls, a)[R(s'ls, a) + V(s')].
This led us to the following formulation in which we exploit the approximate value
function 11 to provide "advice" to the LP optimizer about which constraints should
be tight and which ones should be loose. Consider a state s in L. We can group
the actions available in s into three groups: (a) the "optimal" action a* == 1f(s)
chosen by the approximate policy it, (b) other actions that are tied for optimum
(denoted by ao), and (c) actions that are sub-optimal (denoted by a_). We have
three different constraint equations, one for each type of action:
minimize
L[u(s, a*)
+ v(s, a*)] + LY(s, ao) + L z(s, a_)
s
subject to 17(s)
s,ao
+ u(s, a*) - v(s, a*)
s,a_
== L
P(s'ls, a*)[R(s'ls, a*)
+ V(s')]
s'
17(8) + y(s, ao) ~ L P(s'ls, ao)[R(s'ls, ao)
+ V(s')]
s'
17(s) + z(s, a_) ~ L P(s'ls, a_)[R(s'ls, a_)
+ V(s')] + ?
s'
The first constraint requires V(s) to be approximately equal to the backed-up value
of the chosen optimal action a*. The second constraint requires V(s) to be at least
as large as the backed-up value of any alternative optimal actions ao. If V(s) is
too small, it will be penalized, because the slack variable y(s, ao) will be non-zero.
But there is no penalty if V (s) is too large. The main effect of this constraint is
to drive the value of V(s') downward as necessary to satisfy the first constraint on
a*. Finally, the third constraint requires that V(s) be at least ? larger than the
backed-up value of all inferior actions a_. If these constraints can be satisfied with
all slack variables u, v, y, and z set to zero, then V satisfies the Bellman equation.
After applying the kernel trick and introducing the regularization objective, we
obtain the following Bellman formulation:
minimize
~ aj + C (,~_ u(s, a*) + v(s, a*) + y(s, ao) + z(s, a_))
subject to
~a.j [K(X(Sj),X(S)) J
LP(s'ls,a*)K(X(Sj),X(S'))] +
s'
u(s, a*) - v(s, a*) == L P(s'ls, a*)R(s'ls, a*)
~aj [K(X(Sj),X(S)) J
8'
LP(s'ls,ao)K(X(Sj),X(S'))] +y(s,ao)
~
~ LP(s'ls,ao)R(s'ls,ao)
s'
~O:j [K(X(Sj),X(S)) -
LP(S'IS,a_)K(X(Sj),X(S'))] +z(s,a_)
~
3
~ LP(s'ls,a_)R(s'ls,a_)
+?
8/
Formulation 3: Advantage Learning. The third formulation focuses on the
minimal constraints that must be satisfied to ensure that the greedy policy computed from V will be identical to the greedy policy computed from V (cf. Utgoff
& Saxena, 1987). Specifically, we require that the backed up value of the optimal
action a* be greater than the backed up values of all other actions a.
minimize
L
u(s,a*,a)
s,a*,a
subject to L P(s'ls, a*)[R(s'ls, a*)
+ V(s')] + u(s, a*, a)
8/
~ LP(s!ls,a)[R(s!ls,a)
+ V(s/)] +?
s/
There is one constraint and one slack variable u(s, a*, a) for every action executable
in state s except for the chosen optimal action a* = i"(s). The backed-up value of
a* must have an advantage of at least ? over any other action a, even other actions
that, according to V, are just as good as a*. After applying the kernel trick and
incorporating the complexity penalty, this becomes
minimize
Laj+C L
u(s,a*,a)
s,a*,a
j
subject to Laj L[P(s'ls,a*) -P(s'ls,a)]K(X(sj),X(s')) +u(s,a*,a) ~
j
s/
L P(s'ls, a)R(s'ls, a) - L P(s'ls, a*)R(s'ls, a*) + ?
s/
s/
Of course each of these formulations can easily be modified to incorporate a discount
factor for discounted cumulative reward.
4
Experimental Results
To compare these three formulations, we generated a set of 10 random maze problems as follows. In a 100 by 100 maze, the agent starts in a randomly-chosen square
in the left column, (0, y). Three actions are available in every state, east, northeast,
and southeast, which deterministically move the agent one square in the indicated
direction. The maze is filled with 3000 rewards (each of value -5) generated randomly from a mixture of a uniform distribution (with probability 0.20) and five 2-D
gaussians (each with probability 0.16) centered at (80,20), (80,60), (40,20), (40,80),
and (20,50) with variance 10 in each dimension. Multiple rewards generated for
a single state are accumulated. In addition, in column 99, terminal rewards are
generated according to a distribution that varies from -5 to +15 with minima at
(99,0), (99,40), and (99,80) and maxima at (99,20) and (99,60).
Figure 1 shows one of the generated mazes. These maze problems are surprisingly hard because unlike "traditional" mazes, they contain no walls. In traditional
n;tazes, the walls tend to guide the agent to the goal states by reducing what would
be a 2-D random walk to a random walk of lower dimension (e.g., 1-D along narrow
halls).
100
Rewards
90
-5
-10
80
-15
)(
-20
3IE
0
+
70
CZl
CZl
(1)
~
~
Vi
60
Vi
.s
50
Vi
40
bJJ
~
~
~
.?
(1)
~
30
20
10
10
20
30
40
50
60
70
80
90
100
Figure 1: Example randomly-generated maze. Agent enters at left edge and exits
at right edge.
We applied the three LP formulations in an incremental-batch method as shown
in Table 1. The LPs were solved using the CPLEX package from ILOG. The V
giving the best performance on the starting states in So over the 20 iterations
was saved and evaluated over all 100 possible starting states to obtain a measure of
generalization. The values of C and a were determined by evaluating generalization
on a holdout set of 3 start states: (0,30), (0,50), and (0,70). Experimentation showed
that C = 100,000 worked well for all three methods. We tuned 0- 2 separately for
each problem using values of 5, 10, 20, 40, 60, 80, 120, and 160; larger values
were preferred in case of ties, since they give better generalization. The results are
summarized in Figure 2.
The figure shows that the three methods give essentially identical performance, and
that after 3 examples, all three methods have a regret per start state of about
2 units, which is less than the cost of a single -5 penalty. However, the three
formulations differ in their ease of training and in the information they require.
Table 2 compares training performance in terms of (a) the CPU time required for
training, (b) the number of support vectors constructed, (c) the number of states in
which V prefers a tied-optimal action over the action chosen by n-, (d) the number
of states in which V prefers an inferior action, and (e) the number of iterations
performed after the best-performing iteration on the training set. A high score
on this last measure indicates that the learning algorithm is not converging well,
even though it may momentarily attain a good fit to the data. By virtually every
measure, the advantage formulation scores better. It requires much less CPU time
to train, finds substantially fewer support vectors, finds function approximators that
give better fit to .the data, and tends to converge better. In addition, the advantage?
Table 1:__ Incremental Batch Reinforcement Learning
Repeat 20 times:
For each start state So E 80 do
Generate 16 f-greedy trajectories using V
Record all transitions and rewards to build MDP model
Solve M via value "iteration to obtain V and 7r
if
L=0
For each start state 80 E 80 do
Generate trajectory according to -IT
Add to L all states visited along this trajectory
Apply LP method to L, V, and 7r to find new V
Perform Monte Carlo rollouts using greedy policy for V to evaluate each possible start state
Report total value of all start states.
Table 2: Measures of the quality of the training process (average over 10 MDPs)
180 1=
Sup
Bel
Adv
CPU
37.5
30.4
11.7
#SV
29.5
40.9
17.2
Sup
Bel
Adv
CPU
433.2
208.0
74.5
#SV
105.5
82.4
58.6
180 1=
#tie
70.5
62.0
46.7
1801 =
1
#tie
22.4
18.8
19.4
#bad
0.7
0.9
0.2
#iter
5.6
5.9
1.6
CPU
190.7
92.7
38.4
#SV
54.3
51.1
39.6
#bad
3.0
2.2
0.6
#iter
10.5
3.3
4.0
CPU
789.1
379.1
122.4
#SV
117.2
145.7
74.0
2
#tie
49.8
47.9
29.1
#bad
1.9
0.4
1.4
#iter
7.3
8.2
2.0
#bad
3.3
1.8
3.2
#iter
9.6
7.3
2.8
1801 =4
3
#tie
90.5
75.2
51.9
and Bellman formulations do not require the value of V, but only -fr. This makes
them suitable for learning to imitate a human-supplied policy.
5
Conclusions
This paper has presented three formulations. of batch value function approximation
by exploiting the power of linear programming to express a variety of constraints
and borrowing the kernel trick from support vector machines. All three formulations
were able to learn and generalize well on difficult synthetic maze problems. The
advantage formulation is easier and more reliable to train, probably because it places
fewer constraints on the value function approximation. Hence, we are now applying
the advantage formulation to combinatorial optimization problems in scheduling
and protein structure determination~
Acknowledgments
The authors gratefully acknowledge the support of AFOSR under contract F4962098-1-0375, and the NSF under grants IRl-9626584, I1S-0083292, 1TR-5710001197,
and E1A-9818414. We thank Valentina Zubek and Adam Ashenfelter for their
careful reading of the paper.
1200
~
u
1000
~
0
0.
~
.sa
800
0
B
'"d
Q)
a
a
600
0
~
......
Q.)
l-I
b1)
400
Q)
l-I
~
(5
~
t
200
0
0
2
3
4
5
Number of Starting States
Figure 2: Comparison of the total regret (optimal total reward - attained total
reward) summed over all 100 starting states for the three formulations as a function
of the number of start states in So. The three error bars represent the performance
of the supervised, Bellman, and advantage formulations (left-to-right). The bars
plot the 25th, 50th, and 75th percentiles computed over 10 randomly generated
mazes. Average optimal total reward on these problems is 1306. The random
policy receives a total reward of -14,475.
References
Baird, L. C. (1993). Advantage updating. Tech. rep. 93-1146, Wright-Patterson
AFB.
Moll, R., Barto, A. G., Perkins, T. J., & Sutton, R. S. (1999). Learning instanceindependent value functions to enhance local search. NIPS-II, 1017-1023.
Tesauro, G. (1995). Temporal difference learning and TD-Gammon. CACM, 28(3),
58-68.
Utgoff, P. E., & Saxena, S. (1987). Learning a preference predicate. In ICML-87,
115-121.
Vapnik, V. (2000). The Nature of Statistical Learning Theory, 2nd Ed. Springer.
Zhang, W., & Dietterich, T. G. (1995). A reinforcement learning approach to jobshop scheduling. In IJCAI95, 1114-1120.
| 2116 |@word norm:4 nd:1 seek:4 tr:1 series:1 contains:1 score:2 tuned:1 existing:1 written:1 must:3 plot:1 greedy:4 fewer:2 imitate:1 record:1 preference:1 zhang:2 five:1 mathematical:2 along:4 constructed:1 fitting:2 combine:1 introduce:2 expected:2 themselves:1 terminal:3 bellman:10 discounted:1 td:2 little:1 cpu:6 becomes:1 zubek:1 what:1 substantially:1 maxa:1 temporal:1 every:3 saxena:2 tie:5 unit:1 ly:1 grant:1 local:1 treat:1 tends:1 sutton:2 path:2 approximately:1 studied:1 ease:1 acknowledgment:1 regret:2 implement:3 episodic:1 attain:1 gammon:2 protein:1 scheduling:7 applying:3 deterministic:1 backed:7 starting:6 l:35 searching:2 programming:8 trick:7 updating:1 enters:1 solved:1 momentarily:1 adv:2 complexity:3 utgoff:2 reward:12 esi:1 solving:2 tight:1 patterson:1 exit:1 easily:2 represented:1 train:3 distinct:1 monte:1 cacm:1 larger:2 solve:2 ability:1 advantage:11 product:4 fr:1 combining:2 lookahead:1 exploiting:1 optimum:1 produce:1 incremental:2 executing:2 adam:1 derive:1 fixing:1 received:1 sa:1 c:2 differ:1 direction:1 saved:1 exploration:1 centered:1 human:1 require:3 ao:13 generalization:4 wall:2 preliminary:1 hall:1 wright:1 exp:1 substituting:1 driving:1 optimizer:1 smallest:1 combinatorial:2 visited:5 southeast:1 weighted:2 minimization:1 sensor:2 gaussian:1 modified:1 rather:1 arose:1 ej:4 shuttle:3 barto:2 focus:1 backgammon:2 indicates:1 tech:1 accumulated:1 unlikely:1 borrowing:1 among:1 denoted:2 constrained:2 summed:1 equal:1 once:1 construct:2 identical:2 icml:1 report:1 randomly:4 individual:1 rollouts:1 cplex:1 attempt:1 adjust:1 introduces:1 mixture:1 tgd:1 edge:2 partial:1 necessary:1 experience:1 filled:1 euclidean:1 walk:2 desired:1 minimal:1 fitted:1 column:2 maximization:1 cost:1 introducing:1 deviation:2 uniform:1 i1s:1 successful:1 northeast:1 predicate:1 too:2 reported:1 varies:1 sv:4 synthetic:2 ie:1 contract:1 enhance:1 squared:1 satisfied:2 return:1 summarized:1 availability:1 baird:2 oregon:2 satisfy:1 vi:3 performed:1 sup:2 reached:1 start:9 iwi:1 minimize:13 square:2 variance:1 yield:1 identify:1 yes:1 generalize:5 carlo:1 trajectory:4 drive:1 ed:1 con:1 holdout:1 attained:1 supervised:4 afb:1 formulation:31 evaluated:1 though:1 just:1 receives:1 replacing:2 ei:1 irl:1 quality:1 aj:4 indicated:1 mdp:4 dietterich:2 effect:1 contain:1 hence:3 regularization:1 entering:1 during:4 game:1 inferior:2 percentile:1 generalized:1 complete:1 executable:1 insensitive:1 corvallis:2 similarly:1 gratefully:1 had:1 dot:3 robot:1 add:1 showed:1 driven:1 tesauro:2 bjj:1 rep:1 approximators:1 minimum:1 greater:1 employed:3 converge:1 ii:1 multiple:1 determination:1 visit:1 converging:1 regression:1 essentially:1 iteration:6 kernel:12 represent:5 sometimes:1 laj:3 addition:2 separately:1 unlike:2 probably:1 subject:9 tend:1 virtually:2 call:1 easy:1 variety:2 fit:4 moll:2 jobshop:1 inner:1 idea:2 valentina:1 tradeoff:1 penalty:4 action:23 prefers:2 discount:1 ang:1 generate:2 sl:2 supplied:1 nsf:1 per:1 discrete:2 write:1 express:3 group:2 key:1 iter:4 rewriting:1 sum:2 package:1 parameterized:3 place:1 decision:1 quadratic:2 constraint:15 perkins:2 worked:1 x2:2 performing:1 department:2 according:3 representable:2 remain:1 lp:12 equation:8 resource:1 previously:1 describing:1 slack:9 loose:1 needed:1 studying:1 available:3 operation:1 gaussians:1 experimentation:1 apply:4 generic:1 batch:4 alternative:1 thomas:1 ensure:2 cf:1 exploit:2 giving:1 build:1 move:3 objective:5 primary:1 traditional:2 gradient:2 thank:1 reason:4 minimizing:1 difficult:4 negative:1 rise:1 policy:22 perform:1 ilog:1 markov:1 finite:1 acknowledge:1 situation:1 ever:1 introduced:1 required:1 bel:2 orst:2 learned:1 narrow:1 nip:1 ajx:3 address:1 able:4 bar:2 below:1 reading:2 thail:1 program:1 including:1 reliable:1 power:1 critical:1 suitable:1 mdps:3 f4962098:1 afosr:1 loss:1 approximator:7 agent:5 sufficient:1 consistent:1 course:2 penalized:1 surprisingly:1 last:1 repeat:1 infeasible:1 side:1 guide:1 absolute:2 dimension:2 transition:2 cumulative:1 maze:10 evaluating:1 author:1 reinforcement:3 ashenfelter:1 sj:13 approximate:2 preferred:2 tains:1 b1:1 a_:14 search:4 continuous:2 table:4 learn:2 nature:1 excellent:1 cl:1 domain:1 main:1 arise:1 x1:1 advice:1 ff:1 board:1 experienced:1 position:1 sub:1 deterministically:1 xl:1 lie:1 invocation:1 tied:2 third:3 formula:1 bad:5 svm:1 incorporating:1 vapnik:2 downward:1 easier:2 generalizing:1 led:1 prevents:1 contained:1 applies:2 springer:1 satisfies:1 goal:3 careful:1 ajk:2 hard:1 specifically:2 except:1 reducing:1 determined:1 total:6 experimental:1 xin:1 attempted:1 east:1 formally:1 support:10 incorporate:1 evaluate:1 |
1,226 | 2,117 | Effective size of receptive fields of inferior
temporal visual cortex neurons in natural scenes
Thomas P. Trappenberg
Dalhousie University
Faculty of Computer Science
5060 University Avenue, Halifax B3H 1W5, Canada
[email protected]
Edmund T. Rolls and Simon M. Stringer
University of Oxford,
Centre for Computational Neuroscience,
Department of Experimental Psychology,
South Parks Road, Oxford OX1 3UD, UK
edmund.rolls,[email protected]
Abstract
Inferior temporal cortex (IT) neurons have large receptive fields when a
single effective object stimulus is shown against a blank background, but
have much smaller receptive fields when the object is placed in a natural
scene. Thus, translation invariant object recognition is reduced in natural
scenes, and this may help object selection. We describe a model which
accounts for this by competition within an attractor in which the neurons
are tuned to different objects in the scene, and the fovea has a higher
cortical magnification factor than the peripheral visual field. Furthermore, we show that top-down object bias can increase the receptive field
size, facilitating object search in complex visual scenes, and providing a
model of object-based attention. The model leads to the prediction that
introduction of a second object into a scene with blank background will
reduce the receptive field size to values that depend on the closeness of
the second object to the target stimulus. We suggest that mechanisms of
this type enable the output of IT to be primarily about one object, so that
the areas that receive from IT can select the object as a potential target
for action.
1 Introduction
Neurons in the macaque inferior temporal visual cortex (IT) that respond to objects or faces
have large receptive fields when a single object or image is shown on an otherwise blank
screen [1, 2, 3]. The responsiveness of the neurons to their effective stimuli independent of
their position on the retina over many degrees is termed translation invariance. Translation
invariant object recognition is an important property of visual processing, for it potentially
enables the neurons that receive information from the inferior temporal visual cortex to
perform memory operations to determine whether for example the object has been seen
before or is associated with reward independently of where the image was on the retina.
This allows correct generalization over position, so that what is learned when an object is
shown at one position on the retina generalizes correctly to other positions [4].
If more than one object is present on the screen, then there is evidence that the neuron
responds more to the object at the fovea than in the parafoveal region [5, 6]. More recently,
it has been shown that if an object is presented in a natural background (cluttered scene),
and the monkey is searching for the object in order to touch it to obtain a reward, then the
receptive fields are smaller than when the monkey performs the same task with the object
against a blank background [7]. We define the size of a receptive field as twice the distance
from the fovea (the centre of the receptive field) to locations at which the response decreases
to half maximal. An analysis of IT neurons that responded to the target stimulus showed
that the average size of the receptive fields shrinks from approximately 56 degrees in a
blank background to approximately 12 degrees with a complex scene [8]. The responses of
an IT cell with a large receptive field are illustrated in Figure 1A. There the average firing
rates of the cell to an effective stimulus that the monkey had to touch on a touch-screen
to receive reward is shown as a function of the angular distance of the object from the
fovea. The solid line represents the results from experiments with the object placed in a
blank background. This demonstrates the large receptive fields of IT cells that have often
been reported in the literature [3]. In contrast, when the object is placed in a natural scene
(cluttered background), the size of the receptive field is markedly smaller (dashed line).
2 The model
We formalized our understanding of how the dependence of the receptive field size on various conditions could be implemented in the ventral visual processing pathway by developing a neural network model with the components sufficient to produce the above effects.
The model utilizes an attractor network representing the inferior temporal visual cortex,
and a neural input layer with several retinotopically organized modules representing the
visual scene in an earlier visual cortical area such as V4 (see Figure 1B). Each independent
module within ?V4? represents a small part of the visual field and receives input from earlier visual areas represented by an input vector for each possible location which is unique
for each object. Each module was 6 deg in width, matching the size of the objects presented
to the network. For the simulations we chose binary random input vectors representing objects with
components set to ones and the remaining
components
set to zeros.
is the number of nodes in each module and
is the sparseness
of the representation.
The structure labeled ?IT? represents areas of visual association cortex such as the inferior
temporal visual cortex and cortex in the anterior part of the superior temporal sulcus in
which neurons provide distributed representations of faces and objects [9, 3]. The activity
of nodes in this structure are governed by leaky integrator dynamics with time
constant
! "
#
$ !"
%"'&)(+*, %- *
/. 10 * !"'&2(435- 3 0 3 %"'&28
76
9:;'<
=9> (1)
# $
0 of the ? th node is determined by a sigmoidal function from the activation
The firing
rate
0
!
"
CBD "E&FHGDIKJL
+M @A %"N
POQ1R! , where the parameters M and O
@A as
.
represent the gain and the offset, respectively. The constant represents the strength of
the activity-dependent global inhibition simulating the effects of inhibitory interneurons.
The external ?top-down? input vector
produces object-selective inputs, which are
used as the attentional drive when a visual search task is simulated. The strength of this
< =9S>
A.
B.
130
Average firing rate
120
Object bias
blank
background
110
IT
100
90
80
70
natural
background
60
50
0
10
20
30
40
50
60
Distance of gaze from target object
C.
1
V4
Weight factor
0.8
0.6
from cortical
magnification
factor
0.4
0.2
Gaussian
0
0
10
20
30
40
50
60
Visual Input
Eccentricity
Figure 1: A) Average activity of a macaque inferior temporal cortex neuron as a function
of the distance of the object from the fovea recorded in a visual search task when the object
was in a blank or a cluttered natural background. B) Outline of the model used in this study
with an attractor network labelled ?IT? that receives topgraphical organised inputs from
an input neural layer labeled ?V4?. Objects close to the fovea produce stronger inputs to
reflect the higher magnification factor of the visual representation close to the fovea. The
attractor network also receives top-down object-based inputs, to incorporate object-based
attention in a visual search task. C) The modulation factor used to weight inputs to IT from
V4 shown as a function of their distance from the fovea. The values on the solid line are
derived from cortical magnification factors, and were used in the simulations, whereas the
dotted line corresponds to a shifted Gaussian function.
8 9SA:;
. The recognition functionality of this
object bias is modulated by the value of
structure is modeled as an attractor neural network (ANN) with trained memories indexed
by representing particular objects. The memories are formed through Hebbian learning
on sparse patterns,
(2)
8
where A
*
- 8 (
P Q *
P N
S
*
- A76
*
*
- 6 8 A 76 8S(
P
(set to 1in
the simulations below) is a normalization constant that depends
on
the learning rate,
is the sparseness of the training pattern in IT, and
are the
components of the pattern used to train the network. The weights
between the V4
nodes and IT nodes are trained by Hebbian learning of the form
(3)
to produce object representations in IT based on inputs in V4. The normalizing modulation
8 76 8
8
factor
allows the gain of inputs to be modulated as a function of their distance
from the fovea, and depends on the module to which the presynaptic node belongs. The
weight values between V4 and IT support translation invariant object recognition of a single
object in the visual field if the normalization factor is the same for each module and the
model is trained with the objects placed at every possible location in the visual field. The
translation invariance of the weight vectors between each V4 module and the IT nodes is
however explicitly modulated in our model by the module-dependent modulation factor
as indicated in Figure 1B by the width of the lines connecting V4 with IT.
The strength of the foveal module is strongest, and the strength decreases for modules
representing increasing eccentricity. The form of this modulation factor was derived from
the parameterization of the cortical magnification factors given by [10], 1 and is illustrated
in Figure 1C as a solid line. Similar results to the ones presented here can be achieved
with different forms of the modulation factor such as a shifted Gaussian as illustrated by
the dashed line in Figure 1C.
8 6 8S
3 Results
To study the ability of the model to recognize trained objects at various locations relative
to the fovea we tested the network with distorted versions of the objects, and measured
the ?correlation? between the target object and the final state of the attractor network. The
correlation was estimated from the normalized dot product between the target object vector
and the state of the IT network after a fixed amount of time sufficient for the network to
settle into a stable state. The objects were always presented on backgrounds with some
noise (introduced by flipping 2% of the bits in the scene) because the input to IT will
inevitably be noisy under normal conditions of operation. All results shown in the following
represent averages over 10 runs and over all patterns on which the network was trained.
3.1 Receptive fields are large in scenes with blank backgrounds
In the first experiments we placed only one object in the visual scene with different eccentricities relative to the fovea. The results of this simulation are shown in Figure 2A with
the line labeled ?blank background?. The value of the object bias
was set to 0 in
these simulations. Good object retrieval (indicated by large correlations) was found even
when the object was far from the fovea, indicating large IT receptive fields with a blank
background. The reason that any drop is seen in performance as a function of eccentricity is
because flipping 2% of the bits in the V4 modules introduces some noise into the recall process. The results demonstrate that the attractor dynamics can support translation invariant
object recognition even though the weight vectors between V4 and IT are not translation
invariant but are explicitly modulated by the modulation factor
derived from the
cortical magnification factor.
8 9:;
8 76
3.2 The receptive field size is reduced in scenes with complex background
In a second experiment we placed individual objects at all possible locations in the visual
scene representing natural (cluttered) visual scenes. The resulting correlations between
the target pattern and asymptotic IT state are shown in Figure 2A with the line labeled
?natural background?. Many objects in the visual scene are now competing for recognition
by the attractor network, while the objects around the foveal position are enhanced through
the modulation factor derived by the cortical magnification factor. This results in a much
smaller size of the receptive field of IT neurons when measured with objects in natural
1
This parameterization is based on V1 data. However, it was shown that similar forms of the
magnification factor hold also in V4 [11]
A.
B.
Without object bias
blank
background
0.8
Correlation
Correlation
blank
background
1
1
0.6
0.4
0.8
0.6
0.4
natural
background
0.2
0.2
natural background
0
With object bias
0
10
20
30
40
Eccentricity
50
60
0
0
10
20
30
40
50
60
Eccentricity
Figure 2: Correlations as measured by the normalized dot product between the object vector
used to train IT and the state of the IT network after settling into a stable state with a single
object in the visual scene (blank background) or with other trained objects at all possible
locations in the visual scene (natural background). There is no object bias included in the
results shown in
graph A, whereas an object bias is included in the results shown in B with
in the experiments with a natural background and
in the
experiments with a blank background.
8 9SA:;
8 9SA:;
backgrounds.
3.3 Object-based attention increases the receptive field size, facilitating object
search in complex visual scenes
In addition to this major effect of the background on the size of the receptive field, which
parallels and we suggest may account for the physiological findings outlined in the introduction, there is also a dependence of the size of the receptive fields on the level of object
bias provided to the IT network. Examples are shown in Figure 2B where we used an object
bias. The object bias biasses the IT network towards the expected object with a strength determined by the value of
, and has the effect of increasing the size of the receptive
fields in both blank and natural backgrounds (see Figure 2B and compare to Figure 2A).
This models the effect found neurophysiologically [8].2
8 9SA:;
3.4 A second object in a blank background reduces the receptive field size
depending on the distance between the second object and the fovea
In the last set of experiments we placed two objects in an otherwise blank background.
The IT network was biased towards one of the objects designated as the target object (in
for example a visual search task), which was placed on one side of the fovea at different
eccentricities from the fovea. The second object, a distractor object, was placed on the
opposite side of the fovea at a fixed distance of degrees from the fovea. Results for
different values of are shown in 3A. The results indicate that the size of the receptive
field (for the target object) decreases with decreasing distance of the distractor object from
the fovea. The size of the receptive fields (the width at half maximal response) is shown
2
The larger values of
in the experiments with a natural background compared to the
experiments in a blank background reflects the fact that more attention may be needed to find objects
in natural cluttered environments.
in 3B. The size starts to increase linearly with increasing distance of the distractor object
from the fovea until the influence of the distractor on the size of the receptive field levels
off and approaches the value expected for the situation with one object in a visual scene
and a blank background.
A.
B.
70
Size of receptive field
Correlation
1
0.8
d=24
0.6
d=18
d=12
0.4
0.2
0
d=6
0
10
20
30
40
Eccentricity
50
60
60
50
40
30
20
10
5
10
15
20
25
30
35
40
d, distance of distractor from fovea
Figure 3: A) Correlations between the target object and the final state of the IT network
in experiments with two objects in a visual scene with a blank background. The different
curves correspond to different distances of the distractor object from the fovea. The
eccentricity refers to the distance between the target object and the fovea. B) The size of
the receptive field for the target as a function of the distance of the distractor object from
the fovea.
4 Discussion
When single objects are shown in a scene with a blank background, the attractor network
helps neurons to respond to an object with large eccentricities of this object relative to the
fovea of the agent. When the object is presented in a natural scene, other neurons in the
inferior temporal cortex become activated by the other effective stimuli present in the visual
field, and these forward inputs decrease the response of the network to the target stimulus
by a competitive process. The results found fit well with the neurophysiological data, in
that IT operates with almost complete translation invariance when there is only one object
in the scene, and reduces the receptive field size of its neurons when the object is presented
in a cluttered environment.
The model here provides an explanation of the real IT neuronal responses in natural scenes
and makes several predictions that can be explored experimentally. The model is compatible with the models developed by Gustavo Deco and colleagues (see, for example, [12, 13])
while specific simplifications and addition have been made to explore the variations in the
size of receptive fields in IT.
The model accounts for the larger receptive field sizes from the fovea of IT neurons in
natural backgrounds if the target is the object being selected compared to when it is not
selected [8]. The model accounts for this by an effect of top-down bias which simply
biasses the neurons towards particular objects compensating for their decreasing inputs
produced by the decreasing magnification factor modulation with increasing distance from
the fovea. Such object based attention signals could originate in the prefrontal cortex and
could provide the object bias for the inferotemporal cortex [14].
We proposed that the effective variation of the size of the receptive field in the inferior
temporal visual cortex enables the brain areas that receive from this area (including the
orbitofrontal cortex, amygdala, and hippocampal system) to read out the information
correctly from the inferior temporal visual cortex about individual objects, because the
neurons are responding effectively to the object close to the fovea, and respond very
much less to objects away from the fovea.3 This enables, for example, the correct reward
association of an object to be determined by pattern association in the orbitofrontal cortex
or amygdala, because they receive information essentially about the object at the fovea.
Without this shrinkage in the receptive field size, the areas that receive from the inferior
temporal visual cortex would respond to essentially all objects in a visual scene, and would
therefore provide an undecipherable babel of information about all objects present in the
visual scene. It appears that part of the solution to this potential binding problem that
is used by the brain is to limit the size of the receptive fields of inferior temporal cortex
neurons when natural environments are being viewed. The suggestion is that by providing
an output about what is at the fovea in complex scenes, the inferior temporal visual cortex
enables the correct reward association to be looked up in succeeding brain regions, and
then for the object to be selected for action if appropriate. Part of the hypothesis here is
that the coordinates of the object in the visual scene being selected for action are provided
by the position in space to which the gaze is directed [7].
Acknowledgments
This research was supported by the Medical Research Council, grant PG9826105, and by
the MRC Interdisciplinary Research Centre for Cognitive Neuroscience.
References
[1] C. G. Gross, R. Desimone, T. D. Albright, and Schwartz E. L. Inferior temporal cortex and
pattern recognition. Experimental Brain Research, 11:179?201, 1985.
[2] M. J. Tovee, E. T. Rolls, and P. Azzopardi. Translation invariance and the responses of neurons
in the temporal visual cortical areas of primates. Journal of Neurophysiology, 72:1049?1060,
1994.
[3] E. T. Rolls. Functions of the primate temporal lobe cortical visual areas in invariant visual
object and face recognition. Neuron, 27:205? 218, 2000.
[4] E. T. Rolls and A. Treves. Neural Networks and Brain Function. Oxford University Press,
Oxford, 1998.
[5] T. Sato. Interactions of visual stimuli in the receptive fields of inferior temporal neurons in
macaque. Experimental Brain Research, 77:23?30, 1989.
[6] E. T. Rolls and M. J. Tovee. The responses of single neurons in the temporal visual cortical
areas of the macaque when more than one stimulus is present in the visual field. Experimental
Brain Research, 103:409?420, 1995.
[7] E. T. Rolls, B. Webb, and M. C. A. Booth. Responses of inferior temporal cortex neurons to
objects in natural scenes. Society for Neuroscience Abstracts, 26:1331, 2000.
[8] E. T. Rolls, F. Zheng, and N. Aggelopoulos. Responses of inferior temporal cortex neurons to
objects in natural scenes. Society for Neuroscience Abstracts, 27, 2001.
[9] M. C. A. Booth and E. T. Rolls. View-invariant representations of familiar objects by neurons
in the inferior temporal visual cortex. Cerebral Cortex, 8:510?523, 1998.
3
Note that it is possible that a ?spotlight of attention? [15] can be moved away from the fovea,
but at least during normal visual search tasks, the neurons are sensitive to the object at which the
monkey is looking, that is which is on the fovea, as shown by [8]. Thus, spatial modulation of the
responsiveness of neurons at the V4 level can be influenced by location-specific attentional modulations originating, for example, in the posterior parietal cortex, which may be involved in directing
visual spatial attention [15].
[10] B.W. Dow, A.Z. Snyder, R.G. Vautin, and R. Bauer. Magnification factor and receptive field
size in foveal striate cortex of the monkey. Exp. Brain. Res., 44:213:228, 1981.
[11] R. Gattass, A.P.B. Sousa, and E. Covey. Cortical visual areas of the macaque: Possible substrates for pattern recognition mechanisms. Exp. Brain. Res., Supplement 11, 1985.
[12] G. Deco and J. Zihl. Top-down selective visual attention: A neurodynamical approach. Visual
Cognition, 8:119?140, 2001.
[13] E. T. Rolls and G. Deco. Computational neuroscience of vision. Oxford University Press,
Oxford, 2002.
[14] A. Renart, N. Parga, and E. T. Rolls. A recurrent model of the interaction between the prefrontal
cortex and inferior temporal cortex in delay memory tasks. In S.A. Solla, T.K. Leen, and K.-R.
Mueller, editors, Advances in Neural Information Processing Systems. MIT Press, Cambridge
Mass, 2000. in press.
[15] R. Desimone and J. Duncan. Neural mechanisms of selective visual attention. Annual Review
of Neuroscience, 18:193?222, 1995.
| 2117 |@word neurophysiology:1 version:1 faculty:1 stronger:1 simulation:5 lobe:1 solid:3 foveal:3 tuned:1 blank:22 anterior:1 activation:1 enables:4 drop:1 succeeding:1 half:2 selected:4 parameterization:2 provides:1 node:7 location:7 sigmoidal:1 become:1 pathway:1 expected:2 distractor:7 integrator:1 brain:9 compensating:1 decreasing:3 increasing:4 provided:2 mass:1 what:2 monkey:5 developed:1 finding:1 temporal:23 every:1 tovee:2 demonstrates:1 uk:2 schwartz:1 medical:1 grant:1 before:1 limit:1 oxford:6 firing:3 modulation:10 approximately:2 chose:1 twice:1 directed:1 unique:1 acknowledgment:1 area:11 matching:1 road:1 refers:1 suggest:2 close:3 selection:1 influence:1 attention:9 independently:1 cluttered:6 formalized:1 searching:1 variation:2 coordinate:1 target:14 enhanced:1 substrate:1 hypothesis:1 recognition:9 magnification:10 labeled:4 module:11 region:2 solla:1 decrease:4 gross:1 environment:3 reward:5 dynamic:2 trained:6 depend:1 various:2 represented:1 train:2 effective:6 describe:1 larger:2 otherwise:2 ability:1 noisy:1 final:2 interaction:2 maximal:2 product:2 moved:1 competition:1 eccentricity:10 produce:4 object:111 help:2 depending:1 recurrent:1 ac:1 measured:3 implemented:1 c:1 indicate:1 correct:3 functionality:1 enable:1 settle:1 generalization:1 hold:1 around:1 normal:2 exp:2 cognition:1 major:1 ventral:1 council:1 sensitive:1 reflects:1 mit:1 gaussian:3 always:1 shrinkage:1 derived:4 contrast:1 psy:1 mueller:1 dependent:2 originating:1 selective:3 spatial:2 field:42 represents:4 park:1 stimulus:9 primarily:1 retina:3 recognize:1 individual:2 familiar:1 attractor:9 w5:1 interneurons:1 zheng:1 introduces:1 activated:1 parafoveal:1 desimone:2 indexed:1 re:2 dal:1 earlier:2 poq:1 delay:1 reported:1 interdisciplinary:1 v4:14 off:1 gaze:2 connecting:1 reflect:1 recorded:1 deco:3 prefrontal:2 external:1 cognitive:1 account:4 potential:2 explicitly:2 depends:2 view:1 start:1 competitive:1 parallel:1 simon:2 formed:1 roll:11 responded:1 correspond:1 parga:1 produced:1 mrc:1 drive:1 strongest:1 influenced:1 against:2 colleague:1 involved:1 associated:1 gain:2 recall:1 organized:1 appears:1 higher:2 response:9 leen:1 ox:1 shrink:1 though:1 furthermore:1 angular:1 correlation:9 until:1 receives:3 dow:1 touch:3 ox1:1 indicated:2 effect:6 normalized:2 read:1 illustrated:3 during:1 width:3 inferior:19 hippocampal:1 outline:1 tt:1 demonstrate:1 complete:1 performs:1 image:2 recently:1 superior:1 retinotopically:1 cerebral:1 association:4 spotlight:1 cambridge:1 outlined:1 centre:3 had:1 dot:2 stable:2 cortex:28 inhibition:1 inferotemporal:1 posterior:1 showed:1 belongs:1 termed:1 binary:1 responsiveness:2 seen:2 determine:1 ud:1 dashed:2 signal:1 reduces:2 hebbian:2 retrieval:1 prediction:2 renart:1 essentially:2 vision:1 represent:2 normalization:2 achieved:1 cell:3 receive:6 background:35 whereas:2 addition:2 biased:1 vautin:1 markedly:1 south:1 fit:1 psychology:1 competing:1 opposite:1 reduce:1 avenue:1 whether:1 action:3 amount:1 reduced:2 inhibitory:1 shifted:2 dotted:1 neuroscience:6 estimated:1 correctly:2 snyder:1 sulcus:1 v1:1 graph:1 run:1 respond:4 distorted:1 almost:1 utilizes:1 duncan:1 orbitofrontal:2 bit:2 layer:2 simplification:1 annual:1 activity:3 sato:1 strength:5 scene:32 department:1 developing:1 designated:1 peripheral:1 smaller:4 primate:2 invariant:7 mechanism:3 needed:1 generalizes:1 operation:2 away:2 appropriate:1 simulating:1 sousa:1 thomas:1 responding:1 top:5 remaining:1 cbd:1 society:2 flipping:2 looked:1 receptive:36 dependence:2 striate:1 responds:1 fovea:32 stringer:2 distance:15 attentional:2 simulated:1 originate:1 presynaptic:1 reason:1 modeled:1 providing:2 webb:1 potentially:1 perform:1 neuron:27 inevitably:1 parietal:1 situation:1 looking:1 directing:1 canada:1 treves:1 introduced:1 learned:1 macaque:5 below:1 pattern:8 including:1 memory:4 explanation:1 natural:23 settling:1 representing:6 review:1 literature:1 understanding:1 relative:3 asymptotic:1 neurophysiologically:1 babel:1 suggestion:1 organised:1 degree:4 agent:1 sufficient:2 editor:1 translation:9 compatible:1 placed:9 last:1 supported:1 bias:15 side:2 face:3 leaky:1 sparse:1 distributed:1 bauer:1 curve:1 amygdala:2 cortical:11 forward:1 made:1 far:1 deg:1 global:1 search:7 ca:1 complex:5 linearly:1 noise:2 zihl:1 facilitating:2 neuronal:1 screen:3 position:6 governed:1 down:5 specific:2 offset:1 explored:1 physiological:1 closeness:1 evidence:1 normalizing:1 gustavo:1 effectively:1 supplement:1 edmund:2 sparseness:2 booth:2 simply:1 explore:1 neurophysiological:1 visual:52 binding:1 corresponds:1 viewed:1 ann:1 towards:3 labelled:1 experimentally:1 included:2 determined:3 operates:1 invariance:4 experimental:4 albright:1 indicating:1 select:1 neurodynamical:1 support:2 modulated:4 incorporate:1 tested:1 |
1,227 | 2,118 | Cobot: A Social Reinforcement Learning Agent
Charles Lee Isbell, Jr.
Christian R. Shelton
AT&T Labs-Research
Stanford University
Michael Kearns
Satinder Singh
Peter Stone
University of Pennsylvania Syntek Capital AT&T Labs-Research
Abstract
We report on the use of reinforcement learning with Cobot, a software agent
residing in the well-known online community LambdaMOO. Our initial work on
Cobot (Isbell et al.2000) provided him with the ability to collect social statistics
and report them to users. Here we describe an application of RL allowing Cobot
to take proactive actions in this complex social environment, and adapt behavior
from multiple sources of human reward. After 5 months of training, and 3171
reward and punishment events from 254 different LambdaMOO users, Cobot
learned nontrivial preferences for a number of users, modifing his behavior based
on his current state. Here we describe LambdaMOO and the state and action
spaces of Cobot, and report the statistical results of the learning experiment.
1
Introduction
While most applications of reinforcement learning (RL) to date have been to problems
of control, game playing and optimization (Sutton and Barto1998), there has been a recent
handful of applications to human-computer interaction. Such applications present a number
of interesting challenges to RL methodology (such as data sparsity and inevitable violations
of the Markov property). These previous studies focus on systems that encounter human
users one at a time, such as spoken dialogue systems (Singh et al.2000).
In this paper, we report on an RL-based agent for LambdaMOO, a complex, open-ended,
multi-user chat environment, populated by a community of human users with rich and
often enduring social relationships. Our long-term goal is to build an agent who can learn
to perform useful, interesting and entertaining actions in LambdaMOO on the basis of user
feedback. While this is a deliberately ambitious and underspecified goal, we describe here
our implementation, the empirical experiences of our agent so far, and some of the lessons
we have learned about this challenging domain.
In previous work (Isbell et al.2000), we developed the software agent Cobot, who interacted
in various ways with LambdaMOO users. Cobot had two primary functions. First, Cobot
gathered ?social statistics? (e.g. how frequently and in what ways users interacted with
one another), and provided summaries of these statistics as a service. Second, Cobot had
rudimentary chatting abilities based on the application of information retrieval methods
to large documents. The original Cobot was entirely reactive , in that he never initiated
interaction with human users, but would only respond to their actions. As we documented
in our earlier paper, Cobot proved tremendously popular with LambdaMOO users, setting
the stage for our current efforts.
We modified Cobot to allow him to take certain actions (such as proposing conversation
topics, introducing users, or engaging in common word play routines) under his own initiative. The hope is to build an agent that will eventually take unprompted actions that are
meaningful, useful or amusing to users. Rather than hand-code complex rules specifying
Here we mean ?responding only to human-invoked interaction?, rather than ?non-deliberative?.
Characters in LambdaMOO have gender. Cobot?s description to users indicates that he is male.
when each action is appropriate (rules that would be inaccurate and quickly become stale),
we wanted Cobot to learn the individual and communal preferences of users. Thus, we provided a mechanism for users to reward or punish Cobot, and programmed Cobot to use RL
algorithms to alter his behavior on the basis of this feedback. The application of RL (or any
machine learning methodology) to such an environment presents a number of interesting
domain-specific challenges, including:
Choice of an appropriate state space. To learn how to act in a social environment such
as LambdaMOO, Cobot must represent the salient features. These should include social
information such as which users are present, how experienced they are in LambdaMOO,
how frequently they interact with one another, and so on.
Multiple reward sources. Cobot lives in an environment with multiple, often conflicting
sources of reward from different human users. How to integrate these sources reasonably
is a nontrivial empirical question.
Inconsistency and drift of user rewards and desires. Individual users may be inconsistent in the rewards they provide (even when they implicitly have a fixed set of preferences),
and their preferences may change over time (for example, due to becoming bored or irritated with an action). Even when their rewards are consistent, there can be great temporal
variation in their reward pattern.
Variability in user understanding. There is great variation in users? understanding of
Cobot?s functionality, and the effects of their rewards and punishments.
Data sparsity. Training data is scarce for many reasons, including user fickleness, and the
need to prevent Cobot from generating too much spam in the environment.
Irreproducibility of experiments. As LambdaMOO is a globally distributed community
of human users, it is virtually impossible to replicate experiments taking place there.
We do not have any simple answers (nor do we believe that simple answers exist), but here
we provide a case study of our choices and findings. Our primary findings are:
Inappropriateness of average reward. We found that the average reward that Cobot received over time, the standard measure of success for RL experiments, is an inadequate and
perhaps even inappropriate metric of performance in the LambdaMOO domain. Reasons
include that user preferences are not stationary, but drift as users become habituated or
bored with Cobot?s behavior; and the tendency for satisfied users to stop providing Cobot
with any feedback, positive or negative. Despite the inadequacy of average reward, we are
still able to establish several measures by which Cobot?s RL succeeds, discussed below.
A small set of dedicated ?parents?. While many users provided only a moderate or small
amount of RL training (rewards and punishments) to Cobot, a handful of users did invest
significant time in training him.
Some parents have strong opinions. While many of the users that trained Cobot did
not exhibit clear preferences for any of his actions over the others, some users clearly and
consistently rewarded and punished particular actions over the others.
Cobot learns matching policies. For those users who exhibited clear preferences through
their rewards and punishments, Cobot successfully learned corresponding policies of behavior.
Cobot responds to his dedicated parents. For those users who invested the most training time in Cobot, the observed distribution of his actions is significantly altered by their
presence.
Some preferences depend on state. Although some users for whom we have sufficient
data seem to have preferences that do not depend upon the social state features we constructed for the RL, others do in fact appear to change their preferences depending upon
prevailing social conditions.
The outline for the rest of the paper is as follows. In Section 2, we give brief background
on LambdaMOO. In Section 3, we describe our earlier (non-RL) work on Cobot. Section 4
provides some brief background on RL. In Sections 5, 6 and 7 we describe our implementation of Cobot?s RL action space, reward mechanisms and state features, respectively. Our
primary findings are presented in Section 8, and Section 9 offers conclusions.
2
LambdaMOO
LambdaMOO, founded in 1990 by Pavel Curtis at Xerox PARC, is the oldest continuously
operating MUD, a class of online worlds with roots in text-based multiplayer role-playing
games. MUDs (multi-user dungeons) differ from most chat and gaming systems in their
use of a persistent representation of a virtual world, often created by the participants, who
are represented as characters of their own choosing. LambdaMOO appears as a series of
interconnected rooms, populated by users and objects who may move between them. Each
room provides a shared chat channel, and typically has an elaborate text description that
imbues it with its own ?look and feel.? In addition to speech, users express themselves via
a large collection of verbs, allowing a rich set of simulated actions, and the expression of
emotional states:
(1)
(2)
(3)
(4)
(5)
(6)
Buster is overwhelmed by all these deadlines.
Buster begins to slowly tear his hair out, one strand at a time.
HFh comforts Buster.
HFh [to Buster]: Remember, the mighty oak was once a nut like you.
Buster [to HFh]: Right, but his personal growth was assured. Thanks anyway, though.
Buster feels better now.
Lines (1) and (2) are initiated by verb commands by user Buster, expressing his emotional
state, while lines (3) and (4) are examples of verbs and speech acts, respectively, by HFh.
Lines (5) and (6) are speech and verb acts by Buster. Though there are many standard verbs,
such as the use of the verb comfort in line (3) above, the variety is essentially unlimited,
as players have the ability to create their own verbs.
The rooms and objects in LambdaMOO are created by users themselves, who devise descriptions, and control access by other users. Users can also create objects with verbs that
can be invoked by other players. As last count, the database contains 118,154 objects,
including 4836 active user accounts. LambdaMOO?s long existence and its user-created
nature combine to give it one of the strongest senses of virtual community in the on-line
world. Many users have interacted extensively with each other over many years, and users
are widely acknowledged for their contribution of interesting objects. LambdaMOO is an
attractive environment for experiments in AI (Foner1997; Mauldin1994), including learning. The population is generally curious and technically savvy, and users are interested in
automated objects meant to display some form of intelligence.
3
Cobot
Cobot is a software agent residing in LambdaMOO. Like a human user, he connects via telnet, and from the point of view of the LambdaMOO server, is a user with all the rights and
responsibilities implied. Once actually connected, Cobot wanders into the Living Room,
where he spends most of his time. The Living Room is a central public place, frequented
both by many regulars, and by users new to LambdaMOO. There are several permanent
objects in the Living Room, including a couch with various features and a cuckoo clock.
The Living Room usually has between five and twenty users, and is perpetually busy. Over
a year, Cobot noted over 2.5 million separate events (about one event every eleven seconds)
Previously, we implemented a variety of functionality on Cobot centering around gathering and reporting social statistics. Cobot notes who takes what actions, and on whom.
Cobot can answer queries about these statistics, and describe the similarities and differences between users. He also has a rudimentary chatting ability based on the application
of information retrieval methods to large documents. He can also search the web to answer
specific questions posed to him. A more complete description of Cobot?s abilities, and his
early experiences as a social agent in LambdaMOO, can be found in (Isbell et al.2000).
Our focus here is to make Cobot proactive?i.e., let him take actions under his own
initiative?in a way that is useful, interesting, or pleasing to LambdaMOO users. It is
impossible to program rules anticipating when any given action is appropriate in such a
complex and dynamic environment, so we applied reinforcement learning to adapt directly
from user feedback. We emphasize that Cobot?s original reactive functionality remained
on during the RL experiment. Cobot?s persona is largely due to this original functionality,
and we felt it was most interesting, and even necessary, to add RL work in this context.
Null Action
Topic Change (4)
Roll Call (2)
Social
Commentary
Introductions
Choose to remain silent for this time period.
Introduce a conversational topic. Cobot declares that he wants to discuss sports
or politics, or he utters a sentence from either the sports section or political
section of the Boston Globe.
Initiate a ?roll call,? a common word play routine in LambdaMOO. For example, someone who is tired of Monica Lewinsky may emote ?TIRED OF
LEWINSKY ROLL CALL.? Sympathetic users agree with the roll call. Cobot
takes a recent utterance, and extracts either a single noun, or a verb phrase.
Make a comment describing the current social state of the Living Room, such
as ?It sure is quiet? or ?Everyone here is friendly.? These statements are based
on Cobot?s statistics from recent activity. Several different utterances possible,
but they are treated as a single action for RL purposes.
Introduce two users who have not yet interacted in front of Cobot.
Table 1: The 9 RL actions available to Cobot.
4
RL Background
In RL, problems of decision-making by agents interacting with uncertain environments are
usually modeled as Markov decision processes (MDPs). In the MDP framework, at each
time step the agent senses the state of the environment, and chooses and executes an action
from the set of actions available to it in that state. The agent?s action (and perhaps other
uncontrolled external events) cause a stochastic change in the state of the environment. The
agent receives a (possibly zero) scalar reward from the environment. The agent?s goal is to
choose actions so as to maximize the expected sum of rewards over some time horizon. An
optimal policy is a mapping from states to actions that achieves the agent?s goal.
Many RL algorithms have been developed for learning good approximations to an optimal
policy from the agent?s experience in its environment. At a high level, most algorithms
use this experience to learn value functions (or -values) that map state-action pairs to
the maximal expected sum of reward that can be achieved starting from that state-action
pair. The learned value function is used to choose actions stochastically, so that in each
state, actions with higher value are chosen with higher probability. In addition, many RL
algorithms use some form of function approximation (parametric representations of complex value functions) both to map state-action features to their values and to map states to
distributions over actions (i.e., the policy). See (Sutton and Barto1998) for an extensive
introduction to RL.
In the next sections, we describe the Cobot?s actions, our choice of state features, and how
we dealt with multiple sources of reward. The particular RL algorithm we use is a variant
of (Sutton et al.1999)?s policy gradient algorithm. Its details are beyond the scope of
this paper; however, see (Shelton2000) for details. One aspect of our RL algorithm that
is relevant to understanding our results is that we use a linear function approximator to
store our policy. In other words, for each state feature, we maintain a vector of real-valued
weights indexed by the possible actions. A positive weight for some action means that the
feature increases the probability of taking that action, while a negative weight decreases
the probability. The weight?s magnitude determines the strength of this contribution.
5
Cobot?s RL Actions
To have any hope of learning to behave in a way interesting to LambdaMOO users, Cobot?s
actions must ?make sense? to them, fit in with the social chat-based environment, and
minimize the risk of causing irritation. Conversation, word play, and emoting routines are
among the most common activity in LambdaMOO, so we designed a set of actions along
these lines, as detailed in Table 1. Many of these actions extract an utterance from the
recent conversations, or from a continually changing external source, such as the online
Boston Globe. Thus a single action may cause an infinite variety of behavior by Cobot.
At set time intervals (only every few minutes on average, to minimize spam), Cobot selects
an action to perform from this set according to a distribution determined by the Q-values
in his current state. Any rewards or punishments received before the next RL action are
attributed to the current action, and used to update Cobot?s value functions. It is worth
remembering that Cobot has two different categories of action: those actions taken proactively as a result of the RL, and those actions taken in response to a user?s action towards
Cobot. Some users are certainly aware of the distinction and can easily determine which
actions fall into which category, but other users may occasionally reward or punish Cobot
in response to a reactive action. Such ?erroneous? rewards and punishments act as a source
of noise in the training process.
6
The RL Reward Function
Cobot learns to behave directly from the feedback of LambdaMOO users, any of whom
can reward or punish him. There are both explicit and implicit feedback mechanisms. We
implemented explicit reward and punish verbs on Cobot that LambdaMOO users can
invoke at any time. These verbs give a numerical (positive and negative, respectively)
training signal to Cobot that is the basis of the RL. The signal is attributed as immediate
feedback for the current state and RL action, and ?backed up? to previous states and actions
in accordance with the standard RL algorithms.
There are several standard LambdaMOO verbs that are commonly used to express, sometimes playfully, approval or disapproval. Examples of the former include the verb hug, and
of the latter the verb spank. In the interest of allowing the RL process to integrate naturally
with the LambdaMOO environment, we chose to accept a number of such verbs as implicit
reward and punishment signals for Cobot; however, such implicit feedback is numerically
weaker than the feedback generated by the explicit mechanisms.
One fundamental design choice is whether to learn a single value function for the entire
community, or to learn separate value functions for each user based on individual feedback,
combining the value functions of those present to determine how to act at each moment.
We opted for the latter for three primary reasons.
First, it was clear that for learning to have any hope of success, ths system must represent
who is present at any given moment?different users simply have different personalities and
preferences. We felt that representing which users are present as additional state features
would throw away valuable domain information, as the RL would have to discover on its
own the primacy of user identity. Having separate reward functions for each user is thus a
way of asserting the importance of identity to the learning process.
Second, despite the extremely limited number of training examples available in this domain
per month), learning must be quick and significant. Without a clear sense that their
(
training has some impact on Cobot?s behavior, users will quickly lose interest in providing
feedback. A known challenge for RL is the ?curse of dimensionality,? (i.e. the size of the
state space increases exponentially with the number of state features). By avoiding the need
to represent the presence or absence of roughly 250 users, we are able to maintain a fairly
small state space and so speed up learning.
Third, we (correctly) anticipated the fact that certain users would provide an inordinate
amount of training to Cobot, and we did not want the overall policy followed by Cobot
to be dominated by the preferences of these individuals. By learning separate policies for
each user, and then combining these policies among those users present, we can limit the
impact any single user can have on Cobot?s actions.
7
Cobot?s RL State Features
The decision to maintain and learn separate value functions for each user means that we
can maintain separate state spaces as well, in the hopes of simplifying states and speeding
learning. Cobot can be viewed as running a large number of separate RL processes in
parallel, with each process having a different state space. The state space for a user contains
a number of features containing statistics about that particular user.
LambdaMOO is a social environment, and Cobot is learning to take social actions, so we
felt that his state features should contain information allowing him to gauge social activity
and relationships. Table 2 provides a description of the state features used for RL by Cobot
for each user. Even though we have simplified the state space by partitioning by user, the
state space for a single user remains sufficiently complex to preclude standard table-based
representation of value functions (also, each user?s state space is effectively infinite, as
there are real-valued state features). Thus, linear function approximation is used for each
user?s policy. Cobot?s RL actions are then chosen according to a mixture of the policies of
the users present. We refer the reader to (Shelton2000) for more details on the method by
which policies are learned and combined.
Social Summary
Vector
Mood Vector
Rates Vector
Current Room
Roll Call Vector
Bias
A vector of four numbers: the rate at which the user is producing events; the
rate at which events are being produced that are directed at the user; the percentage of the other users present who are among this user?s ten most frequently interacted-with users (?playmates?); and the percentage of the other
users present for whom this user is among their top ten playmates.
A vector measuring the recent use of eight groups of common verbs (e.g., one
group includes verbs grin and smile). Verbs were grouped according to how
well their usage was correlated.
A vector measuring the rate at which events are produced by those present.
The room where Cobot currently resides.
Indicates if Cobot?s currently saved roll call text has been used before, if someone has done a roll call since the last time Cobot did, and if there has been a
roll call since the last time Cobot grabbed new text.
Each user has one feature that is always ?on?; that is, this bias is always set to
a value of 1. Intuitively, it is the feature indicating the user?s ?presence.?
Table 2: State space of Cobot. Each user has his own state space and value function; the table thus
describes the state space maintained for a generic user.
8
Experimental Procedure and Findings
Cobot has been present in LambdaMOO more or less continuously since September, 1999.
The RL version of Cobot debuted May 10, 2000. Again, Cobot?s various reactive functionality was left intact for the duration of the RL experiment. Cobot is a working system
with real human users, and we wanted to perform the RL experiment in this context. Upon
launching the RL functionality publicly in the Living Room, Cobot logged all RL-related
data (states visited, actions taken, rewards received from each user, parameters of the value
functions, etc.) from May 10 until October 10, 2000. During this time, 63123 RL actions
were taken (in addition, of course, to many more reactive non-RL actions), and 3171 reward and punishment events were received from 254 different users. The findings we now
summarize are based on these extensive logs:
Inappropriateness of average reward. The most standard and obvious sign of successful RL would
be an increase in the average reward over time. Instead, as shown in Figure 1a, the average cumulative reward received by Cobot actually goes down. However, rather than indicating that users are
becoming more dissatisfied as Cobot learns, the decay in reward reveals some peculiarities of human
feedback in such an open-ended environment. There are at least two difficulties with average cumulative reward in an environment of human users. The first is that humans are fickle, and their tastes
and preferences may drift over time. Indeed, our experiences as users, and with the original reactive
functionality of Cobot, suggest that novelty is highly valued in LambdaMOO. Thus a feature that is
popular and exciting to users when it is introduced may eventually become an irritant (there are many
examples of this phenomenon). In RL terminology, we do not have a fixed, consistent reward function, and thus we are always learning a moving target. While difficult to quantify in such a complex
environment, this phenomenon is sufficiently prevalent in LambdaMOO to cast serious doubts on the
use of average cumulative reward as the primary measure of performance.
The second and related difficulty is that even when users do maintain relatively fixed preferences,
they tend to give Cobot less feedback of either type (reward or punishment) as he manages to learn
their preferences accurately. Simply put, once Cobot seems to be behaving as they wish, users feel
no need to continually provide reward for his ?correct? actions or to punish him for the occasional
?mistake.? This reward pattern is in contrast to typical RL applications, where there is an automated
and indefatigable reward source. Strong empirical evidence for this second phenomenon is provided
by User M and User S. These two users were among Cobot?s most dedicated trainers, each had strong
preferences for certain actions, and Cobot learned to strongly modify his behavior in their presence to
match their preferences. Nevertheless, both users tended to provide less frequent feedback to Cobot
as the experiment progressed, as shown in Figure 1a. We conclude that there are serious conceptual
difficulties with the use of average cumulative reward in such a human-centric application of RL, and
that alternative measures must be investigated, which we do below.
A small set of dedicated ?parents.? Among the 254 users who gave at least one reward or punishment event to Cobot, 218 gave 20 or fewer, while 15 gave 50 or more. Thus, we found that while
many users exhibited a passing interest in training Cobot, there was a small group that was willing to
invest nontrivial time and effort in teaching Cobot their preferences. In particular, User M and User
S, generated 594 and 69 rewards and punishments events, respectively.
By ?event?, we simply mean an RL action that received some feedback. The actual absolute
User O
User B
User C
User P
Roll Call. User O appears to especially dislike roll call actions when there have
been repeated roll calls and/or Cobot is repeating the same roll calls.
Rates. The overall rate of events being generating has slightly more relevance
than that of the rate of events being generated just by User O.
Social Summary. User B is effected by the presence of his friends. Not shown
here are other Social Summary features (deviating about 6 degrees). It appears
that User B is more likely to ignore Cobot when he is with many friends.
Roll Call. User C appears to have strong preferences about Cobot?s behavior
when a ?roll call party? is in progress (i.e., everyone is generating roll calls).
Room. User P would follow Cobot to his home, where he is generally alone, and
has trained him there. He appears to have different preferences for Cobot under
those circumstances.
Table 3: Relevant features for users with non-uniform policies. Several of our top users had some
features that deviated from their bias feature. The second column indicates the number of degrees
between the weight vectors for those features and the weight vectors for the bias feature. We have
only included features that deviated by more than 10 degrees. For the users above the double line,
we have included only features whose weights had a length greater than 0.2. Each of these users had
bias weights of length greater than 1. For those below the line, we have included only features with
a length greater than 0.1 (these all had bias weights of length much less than 1).
Some parents have strong opinions. For the vast majority of users who participated in the RL
training of Cobot, the policy learned was quite close to the uniform distribution. Quantification of
this statement is somewhat complex, since policies are dependent on state. However, we observed
that for most users the learned policy?s dependence on state was weak, and the resulting distribution
near uniform (though there are interesting and notable exceptions, as we shall see below). This
result is perhaps to be expected: most users provided too little feedback for Cobot to detect strong
preferences, and may not have been exhibiting strong and consistent preferences in the feedback
they did provide. However, there was again a small group of users for whom a highly non-uniform
policy was learned. In particular, for Users M and S mentioned above, the resulting policies were
relatively independent of state and their entropies were 0.03 and 1.93, respectively. (The entropy of
the uniform distribution over the actions is 2.2.) Several other users also exhibited less dramatic but
still non-uniform distributions. User M seemed to have a strong preference for roll call actions, with
the learned policy selecting these with probability 0.99, while User S preferred social commentary
actions, with the learned policy giving them probability 0.38. (Each action in the uniform distribution
is given weight 1/9 = 0.11.)
Cobot learns matching policies. In Figure 1b, we demonstrate that the policy learned by Cobot
for User M does in fact reflect the empirical pattern of rewards received over time. Similar results
obtain for User S, not shown here. Thus, repeated feedback given to Cobot for a non-uniform set of
preferences clearly pays off in a corresponding policy.
Cobot responds to his dedicated parents. The policies learned by Cobot for users can have strong
impact on the empirical distribution of actions he actually ends up taking in LambdaMOO. For User
M, we find that his presence causes a significant shift towards his preferences. In other words, Cobot
does his best to ?please? these dedicated trainers whenever they arrive in the Living Room, and
returns to a more uniform policy upon their departure.
Some preferences depend on state. Finally, we show that the policies learned by Cobot sometimes
depend upon the features Cobot maintains in his state. We use two facts about the RL weights
(described in Section 4) maintained by Cobot to determine which features are relevant for a given
user. First, we note that by construction, the RL weights learned for the bias feature described in
Table 2 represent the user?s preferences independent of state (since this feature is always on whenever
the user is present). Second, we note that because we initialized all weights to 0, only features with
non-zero weights will contribute to the policy that Cobot uses. Thus, we can determine that a feature
is relevant for a user if that feature?s weight vector is far from that user?s bias feature weight vector,
and from the all-zero vector. For our purposes, we have used (1) the normalized inner product (the
cosine of the angle between two vectors) as a measure of a feature?s distance from the bias feature,
and (2) a feature?s weight vector length to determine if it is away from zero. These measures show that
for most users, Cobot learned a state-independent policy (e.g., User M prefers roll calls); however, as
we can see in Table 3, Cobot has learned a policy for some users that depends upon state.
numerical reward received may be larger or smaller than 1 at any time, as implicit rewards provide
fractional reward, and the user may repeatedly reward or punish an action, with the feedback being
summed. For example, the total absolute value of rewards and punishments provided by User M was
607.63 over 594 feedback events.
Average Cumulative Reward per Timestep
Rewards / Policy / Empirical Distribution Comparision for User M
0.2
1
reward all users
abs reward all users
reward user M
abs reward user M
reward user S
abs reward user S
0.18
0.16
rewards
policy
empirical
0.8
0.14
0.6
0.4
change
Reward
0.12
0.1
0.2
0.08
0.06
0
0.04
?0.2
0.02
0
0
1
2
3
Time
4
5
?0.4
6
4
x 10
1
2
3
4
5
action
6
7
8
9
Figure 1: a) Average Cumulative Reward Over Time. b) Rewards received, policy learned, and
effect on actions for User M. Figure a) shows that average cumulative reward decreases over time,
for both total and absolute reward; however, Figure b shows that proper learning is taking place. For
each of the RL actions, three quantities are shown. The blue bars (left) show the average reward
given by User M for each action (the average reward given by User M across all actions has been
subtracted off to indicate relative preferences). The yellow bars (middle) show the policy learned by
Cobot for User M (the probability assigned to each action in the uniform distribution (1/9) has been
subtracted off). The red bars (right) show the empirical frequency with which each action was taken
in the presence of User M (minus the empirical frequency with which that action was taken by Cobot
over all time steps). These bars indicate the extent to which the presence of User M biases Cobot?s
behavior towards M?s preferences. We see that the policy learned by Cobot for User M aligns nicely
with the preferences expressed by M and that Cobot?s behavior shifts strongly towards the learned
policy for User M whenever M is present. To go beyond a qualitative visual analysis, we have defined
a metric that measures the extent to which two rankings of actions agree, while taking into account
that some actions are extremely close in the each ranking. The details are beyond the scope of the
paper, but the agreement between the action rankings shown here are in near-perfect agreement by
this measure. Similar results obtain for User S.
9
Conclusions
We have reported on our efforts to apply reinforcement learning in a complex human online
social environment where many of the standard assumptions (stationary rewards, Markovian behavior, appropriateness of average reward) are clearly violated. We feel that the
results obtained with Cobot so far are compelling, and offer promise for the application
of RL in such open-ended social settings. Cobot continues to take RL actions and receive
rewards and punishments from LambdaMOO users, and we plan to continue and embellish
this work as part of our overall efforts on Cobot.
References
Foner, L. (1997). Entertaining Agents: a Sociological Case Study. In Proceedings of the First
International Conference on Autonomous Agents.
Isbell, C. L., Kearns, M., Kormann, D., Singh, S., and Stone, P. (2000). Cobot in LambdaMOO: A
Social Statistics Agent. To appear in Proceedings of AAAI-2000.
Mauldin, M. (1994). Chatterbots, TinyMUDs, and the Turing Test: Entering the Loebner Prize
Competition. In Proceedings of the Twelfth National Conference on Artificial Intelligence.
Shelton, C. R. (2000). Balancing Multiple Sources of Reward in Reinforcement Learning. Submitted
for publication in Neural Information Processing Systems-2000.
Singh, S., Kearns, M., Littman, D., and Walker, M. (2000). Empirical Evaluation of a Reinforcement
Learning Dialogue System. To appear in Proceedings of AAAI-2000.
Sutton, R. S. and Barto, A. G. (1998). Reinforcement Learning: An Introduction. MIT Press, Cambridge, MA.
Sutton, R. S., McAllester, D., Singh, S., and Mansour, Y. (1999). Policy gradient methods for reinforcement learning with function approximation. In Neural Information Processing Systems1999.
| 2118 |@word middle:1 version:1 seems:1 replicate:1 twelfth:1 open:3 willing:1 simplifying:1 pavel:1 dramatic:1 minus:1 moment:2 initial:1 series:1 contains:2 selecting:1 document:2 current:7 yet:1 must:5 numerical:2 entertaining:2 eleven:1 christian:1 wanted:2 designed:1 update:1 stationary:2 intelligence:2 fewer:1 alone:1 oldest:1 prize:1 provides:3 contribute:1 preference:30 launching:1 oak:1 five:1 along:1 constructed:1 become:3 initiative:2 persistent:1 qualitative:1 combine:1 introduce:2 indeed:1 expected:3 roughly:1 themselves:2 frequently:3 nor:1 multi:2 behavior:12 approval:1 globally:1 actual:1 curse:1 inappropriate:1 preclude:1 little:1 provided:7 begin:1 discover:1 null:1 what:2 spends:1 developed:2 proposing:1 spoken:1 finding:5 ended:3 temporal:1 remember:1 every:2 act:5 friendly:1 growth:1 control:2 partitioning:1 appear:3 producing:1 continually:2 positive:3 service:1 before:2 accordance:1 modify:1 limit:1 mistake:1 sutton:5 despite:2 initiated:2 punish:6 inordinate:1 becoming:2 chose:1 collect:1 challenging:1 specifying:1 someone:2 programmed:1 limited:1 directed:1 procedure:1 empirical:10 significantly:1 matching:2 word:5 regular:1 suggest:1 close:2 put:1 context:2 impossible:2 risk:1 map:3 quick:1 backed:1 go:2 starting:1 duration:1 rule:3 his:25 population:1 anyway:1 variation:2 autonomous:1 feel:4 target:1 play:3 construction:1 user:162 savvy:1 us:1 engaging:1 agreement:2 continues:1 underspecified:1 database:1 dissatisfied:1 observed:2 role:1 connected:1 grabbed:1 decrease:2 buster:8 valuable:1 mentioned:1 environment:20 reward:70 littman:1 dynamic:1 personal:1 trained:2 singh:5 depend:4 parc:1 technically:1 upon:6 basis:3 easily:1 various:3 represented:1 describe:7 couch:1 query:1 artificial:1 choosing:1 whose:1 quite:1 stanford:1 widely:1 posed:1 valued:3 larger:1 ability:5 statistic:8 invested:1 online:4 mood:1 interaction:3 product:1 chatting:2 syntek:1 frequent:1 interconnected:1 maximal:1 relevant:4 causing:1 date:1 combining:2 description:5 competition:1 invest:2 interacted:5 parent:6 double:1 generating:3 perfect:1 object:7 depending:1 friend:2 received:9 progress:1 strong:9 throw:1 implemented:2 indicate:2 quantify:1 differ:1 exhibiting:1 appropriateness:1 saved:1 functionality:7 utters:1 stochastic:1 peculiarity:1 correct:1 human:15 mcallester:1 opinion:2 virtual:2 public:1 hfh:4 around:1 residing:2 sufficiently:2 great:2 mapping:1 scope:2 achieves:1 early:1 purpose:2 lose:1 punished:1 currently:2 visited:1 him:9 grouped:1 create:2 successfully:1 gauge:1 hope:4 mit:1 clearly:3 always:4 modified:1 rather:3 command:1 barto:1 publication:1 focus:2 consistently:1 prevalent:1 indicates:3 political:1 tremendously:1 opted:1 contrast:1 sense:2 detect:1 dependent:1 inaccurate:1 typically:1 entire:1 accept:1 interested:1 selects:1 trainer:2 overall:3 among:6 plan:1 prevailing:1 noun:1 fairly:1 summed:1 once:3 never:1 nicely:1 aware:1 having:2 look:1 progressed:1 inevitable:1 alter:1 anticipated:1 report:4 others:3 serious:2 few:1 national:1 individual:4 deviating:1 connects:1 maintain:5 ab:3 pleasing:1 interest:3 highly:2 evaluation:1 certainly:1 violation:1 male:1 mixture:1 sens:2 necessary:1 experience:5 indexed:1 initialized:1 uncertain:1 column:1 earlier:2 compelling:1 markovian:1 measuring:2 phrase:1 introducing:1 uniform:10 successful:1 inadequate:1 too:2 front:1 reported:1 answer:4 punishment:13 chooses:1 thanks:1 combined:1 fundamental:1 international:1 lee:1 invoke:1 off:3 michael:1 continuously:2 quickly:2 monica:1 again:2 aaai:2 central:1 satisfied:1 containing:1 choose:3 slowly:1 possibly:1 reflect:1 external:2 stochastically:1 dialogue:2 return:1 doubt:1 account:2 busy:1 includes:1 permanent:1 notable:1 ranking:3 proactively:1 depends:1 proactive:2 root:1 view:1 lab:2 responsibility:1 red:1 effected:1 participant:1 parallel:1 maintains:1 contribution:2 mud:2 minimize:2 publicly:1 roll:17 who:14 largely:1 gathered:1 lesson:1 yellow:1 dealt:1 weak:1 accurately:1 produced:2 manages:1 worth:1 executes:1 submitted:1 strongest:1 tended:1 whenever:3 aligns:1 centering:1 frequency:2 obvious:1 naturally:1 attributed:2 stop:1 proved:1 popular:2 conversation:3 fractional:1 dimensionality:1 routine:3 anticipating:1 actually:3 appears:5 centric:1 higher:2 follow:1 methodology:2 response:2 done:1 though:4 strongly:2 just:1 stage:1 implicit:4 clock:1 until:1 hand:1 receives:1 working:1 web:1 gaming:1 chat:4 perhaps:3 stale:1 believe:1 mdp:1 usage:1 effect:2 contain:1 normalized:1 deliberately:1 former:1 assigned:1 entering:1 nut:1 attractive:1 game:2 during:2 please:1 maintained:2 noted:1 cosine:1 stone:2 fickle:1 outline:1 complete:1 demonstrate:1 dedicated:6 rudimentary:2 invoked:2 charles:1 common:4 rl:56 exponentially:1 million:1 discussed:1 he:13 numerically:1 significant:3 expressing:1 refer:1 cambridge:1 ai:1 populated:2 teaching:1 had:7 moving:1 access:1 similarity:1 operating:1 behaving:1 etc:1 add:1 own:7 recent:5 moderate:1 rewarded:1 store:1 certain:3 server:1 occasionally:1 success:2 continue:1 life:1 inconsistency:1 devise:1 commentary:2 tired:2 remembering:1 additional:1 greater:3 somewhat:1 determine:5 maximize:1 period:1 novelty:1 living:7 signal:3 multiple:5 match:1 adapt:2 tear:1 offer:2 long:2 retrieval:2 deadline:1 impact:3 variant:1 hair:1 globe:2 essentially:1 metric:2 circumstance:1 represent:4 sometimes:2 achieved:1 receive:1 background:3 addition:3 want:2 participated:1 interval:1 walker:1 source:9 rest:1 exhibited:3 sure:1 comment:1 cobot:120 tend:1 virtually:1 inconsistent:1 seem:1 smile:1 imbues:1 call:17 curious:1 near:2 presence:8 automated:2 variety:3 fit:1 gave:3 pennsylvania:1 silent:1 inner:1 shift:2 politics:1 whether:1 expression:1 inadequacy:1 effort:4 peter:1 speech:3 passing:1 cause:3 action:78 prefers:1 repeatedly:1 useful:3 generally:2 clear:4 detailed:1 amount:2 repeating:1 extensively:1 ten:2 category:2 documented:1 wanders:1 exist:1 percentage:2 lambdamoo:38 sign:1 per:2 correctly:1 blue:1 shall:1 promise:1 express:2 group:4 salient:1 four:1 terminology:1 nevertheless:1 acknowledged:1 capital:1 changing:1 prevent:1 vast:1 timestep:1 year:2 sum:2 angle:1 turing:1 you:1 respond:1 logged:1 place:3 reporting:1 reader:1 arrive:1 home:1 decision:3 entirely:1 uncontrolled:1 pay:1 followed:1 display:1 deviated:2 activity:3 nontrivial:3 strength:1 declares:1 comparision:1 handful:2 isbell:5 software:3 unlimited:1 felt:3 dominated:1 aspect:1 speed:1 extremely:2 conversational:1 relatively:2 according:3 xerox:1 multiplayer:1 jr:1 remain:1 describes:1 slightly:1 character:2 smaller:1 across:1 making:1 lewinsky:2 intuitively:1 gathering:1 taken:6 agree:2 previously:1 remains:1 discus:1 eventually:2 mechanism:4 count:1 describing:1 initiate:1 end:1 available:3 eight:1 apply:1 occasional:1 away:2 appropriate:3 generic:1 primacy:1 alternative:1 encounter:1 subtracted:2 existence:1 original:4 cuckoo:1 responding:1 personality:1 include:3 running:1 top:2 emotional:2 giving:1 build:2 establish:1 especially:1 implied:1 move:1 question:2 quantity:1 parametric:1 primary:5 dependence:1 responds:2 exhibit:1 gradient:2 quiet:1 september:1 distance:1 separate:7 simulated:1 majority:1 topic:3 whom:5 extent:2 reason:3 code:1 length:5 modeled:1 relationship:2 providing:2 difficult:1 october:1 statement:2 negative:3 implementation:2 design:1 ambitious:1 policy:37 twenty:1 perform:3 allowing:4 proper:1 markov:2 behave:2 immediate:1 variability:1 interacting:1 mansour:1 verb:18 community:5 drift:3 introduced:1 pair:2 cast:1 extensive:2 sentence:1 shelton:2 learned:21 conflicting:1 distinction:1 hug:1 able:2 beyond:3 bar:4 below:4 pattern:3 usually:2 comfort:2 departure:1 sparsity:2 challenge:3 summarize:1 program:1 including:5 everyone:2 event:14 treated:1 difficulty:3 quantification:1 scarce:1 representing:1 altered:1 brief:2 mdps:1 created:3 extract:2 utterance:3 speeding:1 text:4 understanding:3 taste:1 dislike:1 relative:1 sociological:1 interesting:8 approximator:1 integrate:2 agent:19 degree:3 sufficient:1 consistent:3 exciting:1 playing:2 balancing:1 course:1 summary:4 last:3 bias:10 allow:1 weaker:1 fall:1 taking:5 absolute:3 distributed:1 feedback:20 world:3 cumulative:7 rich:2 asserting:1 resides:1 seemed:1 collection:1 reinforcement:9 perpetually:1 commonly:1 spam:2 founded:1 far:3 simplified:1 social:24 party:1 emphasize:1 ignore:1 implicitly:1 preferred:1 satinder:1 active:1 reveals:1 inappropriateness:2 conceptual:1 conclude:1 search:1 table:9 learn:8 reasonably:1 channel:1 nature:1 curtis:1 interact:1 investigated:1 complex:9 domain:5 assured:1 did:5 bored:2 noise:1 repeated:2 elaborate:1 experienced:1 explicit:3 wish:1 third:1 learns:4 minute:1 remained:1 erroneous:1 down:1 specific:2 mauldin:1 habituated:1 decay:1 evidence:1 enduring:1 effectively:1 importance:1 magnitude:1 overwhelmed:1 horizon:1 boston:2 entropy:2 simply:3 likely:1 sympathetic:1 visual:1 desire:1 strand:1 expressed:1 sport:2 scalar:1 gender:1 determines:1 ma:1 month:2 goal:4 identity:2 viewed:1 towards:4 room:13 shared:1 absence:1 change:5 included:3 infinite:2 determined:1 typical:1 kearns:3 total:2 tendency:1 experimental:1 succeeds:1 player:2 meaningful:1 intact:1 indicating:2 exception:1 latter:2 meant:1 reactive:6 phenomenon:3 relevance:1 violated:1 avoiding:1 correlated:1 |
1,228 | 2,119 | Probabilistic principles in unsupervised learning
of visual structure: human data and a model
Shimon Edelman, Benjamin P. Hiles
& Hwajin Yang
Department of Psychology
Cornell University, Ithaca, NY 14853
se37,bph7,hy56 @cornell.edu
Nathan Intrator
Institute for Brain and Neural Systems
Box 1843, Brown University
Providence, RI 02912
Nathan [email protected]
Abstract
To find out how the representations of structured visual objects depend
on the co-occurrence statistics of their constituents, we exposed subjects
to a set of composite images with tight control exerted over (1) the conditional probabilities of the constituent fragments, and (2) the value of Barlow?s criterion of ?suspicious coincidence? (the ratio of joint probability
to the product of marginals). We then compared the part verification response times for various probe/target combinations before and after the
exposure. For composite probes, the speedup was much larger for targets that contained pairs of fragments perfectly predictive of each other,
compared to those that did not. This effect was modulated by the significance of their co-occurrence as estimated by Barlow?s criterion. For
lone-fragment probes, the speedup in all conditions was generally lower
than for composites. These results shed light on the brain?s strategies for
unsupervised acquisition of structural information in vision.
1 Motivation
How does the human visual system decide for which objects it should maintain distinct
and persistent internal representations of the kind typically postulated by theories of object
recognition? Consider, for example, the image shown in Figure 1, left. This image can be
represented as a monolithic hieroglyph, a pair of Chinese characters (which we shall refer
to as and ), a set of strokes, or, trivially, as a collection of pixels. Note that the second
option is only available to a system previously exposed to various combinations of Chinese
characters. Indeed, a principled decision whether to represent this image as ,
or otherwise can only be made on the basis of prior exposure to related images.
According to Barlow?s [1] insight, one useful principle is tallying suspicious coincidences:
two candidate fragments and should be combined into a composite object if the
probability of their joint appearance
is much higher than
, which is
the probability expected in the case of their statistical independence. This criterion may be
compared to the Minimum Description Length (MDL) principle, which has been previously
discussed in the context of object representation [2, 3]. In a simplified form [4], MDL calls
for representing explicitly as a whole if
, just as the principle
of suspicious coincidences does.
While the Barlow/MDL criterion
certainly indicates a suspicious coincidence, there are additional probabilistic considerations that may be used
in setting the degree of association between and . One example is the possi
ble perfect predictability of from and vice versa, as measured by
, then and are perfectly predictive of each
. If
other and should really be coded by a single symbol, whereas the MDL criterion may suggest merely that some association between the representation of and that of be estab
lished. In comparison, if and are not perfectly predictive of each other (
),
there is a case to be made in favor of coding them separately to allow for a maximally
expressive representation, whereas MDL may actually suggest a high degree of association
). In this study we investigated whether the human
(if
visual system uses a criterion based on
alongside MDL while learning (in an unsupervised manner) to represent composite objects.
AB
Figure 1: Left: how many objects are contained in image ? Without prior knowledge, a
reasonable answer, which embodies a holistic bias, should be ?one? (Gestalt effects, which
would suggest two convex ?blobs? [5], are beyond the scope of the present discussion).
Right: in this set of ten images, appears five times as a whole; the other five times
a fragment wholly contained in appears in isolation. This statistical fact provides
grounds for considering to be composite, consisting of two fragments (call the upper
one and the lower one ), because
, but
.
To date, psychophysical explorations of the sensitivity of human subjects to stimulus statistics tended to concentrate on means (and sometimes variances) of the frequency of various
stimuli (e.g., [6]. One recent and notable exception is the work of Saffran et al. [7], who
showed that infants (and adults) can distinguish between ?words? (stable pairs of syllables
that recur in a continuous auditory stimulus stream) and non-words (syllables accidentally
paired with each other, the first of which comes from one ?word? and the second ? from
the following one). Thus, subjects can sense (and act upon) differences in transition probabilities between successive auditory stimuli. This finding has been recently replicated, with
infants as young as 2 months, in the visual sequence domain, using successive presentation
of simple geometric shapes with controlled transition probabilities [8]. Also in the visual
domain, Fiser and Aslin [9] presented subjects with geometrical shapes in various spatial
configurations, and found effects of conditional probabilities of shape co-occurrences, in a
task that required the subjects to decide in each trial which of two simultaneously presented
shapes was more familiar.
The present study was undertaken to investigate the relevance of the various notions of
statistical independence to the unsupervised learning of complex visual stimuli by human
subjects. Our experimental approach differs from that of [9] in several respects. First,
instead of explicitly judging shape familiarity, our subjects had to verify the presence of a
probe shape embedded in a target. This objective task, which produces a pattern of response
times, is arguably better suited to the investigation of internal representations involved in
object recognition than subjective judgment. Second, the estimation of familiarity requires
the subject to access in each trial the representations of all the objects seen in the experi-
ment; in our task, each trial involved just two objects (the probe and the target), potentially
sharpening the focus of the experimental approach. Third, our experiments tested the pre
, and MDL, or Barlow?s
dictions of two distinct notions of stimulus independence:
ratio.
2 The psychophysical experiments
In two experiments, we presented stimuli composed of characters such as those in Figure 1
to nearly 100 subjects unfamiliar with the Chinese script. The conditional probabilities
of the appearance of individual characters were controlled. The experiments involved two
types of probe conditions: P TYPE=Fragment, or
(with
as the
(with
as
reference condition), and P TYPE=Composite, or
reference). In this notation (see Figure 2, left), and are ?familiar? fragments with controlled minimum conditional probability
, and are novel (low-probability)
fragments.
Each of the two experiments consisted of a baseline phase, followed by training exposure
(unsupervised learning), followed in turn by the test phase (Figure 2, right). In the baseline
and test phases, the subjects had to indicate whether or not the probe was contained in the
target (a task previously used by Palmer [5]). In the intervening training phase, the subjects
merely watched the character triplets presented on the screen; to ensure their attention, the
subjects were asked to note the order in which the characters appeared.
V
ABZ
VW
baseline/test
ABZ
target
reference
mask
probe
A
ABZ
AB
ABZ
4
test
3
2
1
probe
target
probe
Fragment
target
unsupervised training
Composite
Figure 2: Left: illustration of the probe and target composition for the two levels of P TYPE
(Fragment and Composite). For convenience, the various categories of characters that
appeared in the experiment are annotated here by Latin letters: , stand for characters
with controlled
, and stand for characters that
appeared only once throughout an experiment. In experiment 1, the training set was con
structed with
for some pairs, and
for others; in experiment 2,
Barlow?s suspicious coincidence ratio was also controlled. Right top: the structure of a
part verification trial (same for baseline and test phases). The probe stimulus was followed
by the target (each presented for
; a mask was shown before and after the target).
The subject had to indicate whether or not the former was contained in the latter (in this
example, the correct answer is yes). A sequence consisting of 64 trials like this one was
presented twice: before training (baseline phase) and after training (test phase). For ?positive? trials (i.e., probe contained in target), we looked at the S PEEDUP following training,
; negative trials were discarded. Right bottom: the
defined as
structure of a training trial (the training phase, placed between baseline and test, consisted
of 80 such trials). The three components of the stimulus appeared one by one for
to make sure that the subject attended to each, then together for
. The subject was
required to note whether the sequence unfolded in a clockwise or counterclockwise order.
!
The logic behind the psychophysical experiments rested on two premises. First, we knew
from earlier work [5] that a probe is detected faster if it is represented monolithically (that
is, considered to be a good ?object? in the Gestalt sense). Second, we hypothesized that a
composite stimulus would be treated as a monolithic object to the extent that its constituent
characters are predictable from each other, as measured by a high conditional probability,
, and/or by a high suspicious coincidence ratio, . The main prediction following
from these premises is that the S PEEDUP (the difference in response time between baseline
and test phases) for a composite probe should reflect the mutual predictability of the probe?s
constituents in the training set. Thus, our hypothesis ? that statistics of co-occurrence
determine the constituents in terms of which structured objects are represented ? would
be supported if the S PEEDUP turns out to be larger for those composite probes whose
constituents tend to appear together in the training set. The experiments, therefore, hinged
on a comparison of the patterns of response times in the ?positive? trials (in which the
probe actually is embedded in the target; see Figure 2, left) before and after exposure to the
training set.
400
Composite
Fragment
analog of speedup
0.3
speedup, ms
300
200
100
Composite
Fragment
0.2
0.1
0
?0.1
0
0.4
minCP
0.6
0.8
?0.2
0.4
1
0.6
0.8
minCP
1
Figure 3: Left: unsupervised learning of statistically defined structure by human subjects,
). The dependent variable S PEED - UP is defined as the difference in
experiment 1 (
between baseline and test phases (least-squares estimates of means and standard errors,
computed by the LSMEANS option of SAS procedure MIXED [10]). The S PEED - UP for
composite probes (solid line) with
exceeded that in the other conditions by
. Right: the results of a simulation of experiment 1 by a model derived from
about
the one described in [4]. The model was exposed to the same 80 training images as the
human subjects. The difference of reconstruction errors for probe and target served as the
analog of RT; baseline measurements were conducted on half-trained networks.
2.1 Experiment 1
Fourteen subjects, none of them familiar with the Chinese writing system, participated in
this experiment in exchange for course credit. Among the stimuli, two characters
. Alternatively, could
could be paired, in which case we had
be unpaired, with
,
(in this experiment, we held the suspicious
coincidence ratio
constant at ). For the paired
the minimum conditional probability
and the
two characters were perfectly predictable from each other, whereas for the unpaired
, and they were not. In the latter case probably should not be represented
as a whole.
!
As expected, we found the value of S PEED - UP to be strikingly different for composite
probes with
(
) compared to the other three conditions (about
);
see Figure 3, left. A mixed-effects repeated measures analysis of variance (SAS procedure
MIXED [10]) for S PEED - UP revealed a marginal effect of P TYPE (
) and a significant interaction P TYPE
interaction (
).
!
principle: S PEEDUP was generThis behavior conforms to the predictions of the
ally higher for composite probes, and disproportionately higher for composite probes with
. The subjects in experiment 1 proved to be sensitive to the
measure
of independence in learning to associate object fragments together. Note that the suspi
.
cious coincidence ratio was the same in both cases,
over and above the (constant-valued) MDLThus, the visual system is sensitive to
related criterion, according to which the propensity to form a unified representation of two
fragments, and , should be determined by [1, 4].
200
200
150
100
50
0.8
minCP
minCP=0.5
150
100
50
0
0.4
1
250
250
200
200
speedup, ms
speedup, ms
0.6
150
100
50
0
0
r=8.33
250
speedup, ms
speedup, ms
r=1.13
250
0
0.4
5
r
0.6
0.8
minCP
minCP=1.0
1
150
100
50
0
0
10
5
r
10
Figure 4: Human subjects, experiment 2 (
). The effect of
found in experiment 1 was modulated in a complicated fashion by the effect of the suspicious coincidence
ratio (see text for discussion).
2.2 Experiment 2
together.
In the second experiment, we studied the effects of varying both and
Because these two quantities are related (through the Bayes theorem), they cannot be manipulated independently. To accommodate this constraint, some subjects saw two sets of
, in the first sesstimuli, with
and with
and with
sion and other two sets, with
,
in the second session; for other subjects, the complementary combinations were used in
each session. Eighty one subjects unfamiliar with the Chinese script participated in this
experiment for course credit.
The results (Figure 4) showed that S PEEDUP was consistently higher for composite probes.
Thus, the association between probe constituents was strengthened by training in each of
the four conditions. S PEEDUP was also generally higher for the high suspicious coinci , and disproportionately higher for composite probes in the
dence ratio case,
case, indicating a complicated synergy between the two mea,
sures of dependence,
and . A mixed-effects repeated measures analysis of variance (SAS procedure MIXED [10]) for S PEED - UP revealed significant main effects of
) and (
P TYPE (
), as well as
) and
two significant two-way interactions,
(
). There was also a marginal three-way interaction,
P TYPE (
P TYPE (
).
!
!
!
The findings of these two psychophysical experiments can be summarized as follows: (1)
an individual complex visual shape (a Chinese character) is detected faster than a composite stimulus (a pair of such characters) when embedded in a 3-character scene, but this
advantage is narrowed with practice; (2) a composite attains an ?objecthood? status to the
extent that its constituents are predictable from each other, as measured either by the conditional probability,
, or by the suspicious coincidence ratio, ; (3) for composites,
the strongest boost towards objecthood (measured by response speedup following unsuper
is high and is low, or vice versa. The nature of
vised learning) is obtained when
this latter interaction is unclear, and needs further study.
3 An unsupervised learning model and a simulated experiment
The ability of our subjects to construct representations that reflect the probability of cooccurrence of complex shapes has been replicated by a pilot version of an unsupervised
learning model, derived from the work of [4]. The model (Figure 5) is based on the following observation: an auto-association network fed with a sequence of composite images
in which some fragment/location combinations are more likely than others develops a nonuniform spatial distribution of reconstruction errors. Specifically, smaller errors appear in
those locations where the image fragments recur. This information can be used to form a
spatial receptive field for the learning module, while the reconstruction error can signal its
relevance to the current input [11, 12].
In the simplified pilot model, the spatial receptive field (labeled in Figure 5, left, as ?relevance mask?) consists of four weights, one per quadrant:
,
. During
the
,
unsupervised training, the weights are updated by setting
where is the reconstruction error in trial , and and are learning constants. In a
simulation of experiment 1, a separate module with its own four-weight ?receptive field?
was trained for each of the composite stimuli shown to the human subjects. 1 The Euclidean
distance between probe and target representations at the output of the model served as the
analog of response time, allowing us to compare the model?s performance with that of the
humans. We found the same differential effects of
for Fragment and Composite probes in the real and simulated experiments; compare Figure 3, left (humans) with
Figure 3, right (model).
1
The full-fledged model, currently under development, will have a more flexible receptive field
structure, and will incorporate competitive learning among the modules.
input
error
input
adapt
?
erri
relevance
mask (RF)
ensemble of modules
auto?
associator
reconstructed
Figure 5: Left: the functional architecture of a fragment module. The module consists of
two adaptive components: a reconstruction network, and a relevance mask, which assigns
different weights to different input pixels. The mask modulates the input multiplicatively,
determining the module?s receptive field. Given a sequence of images, several such modules working in parallel learn to represent different categories of spatially localized patterns
(fragments) that recur in those images. The reconstruction error serves as an estimate of
the module?s ability to deal with the input ([11, 12]; in the error image, shown on the right,
white corresponds to high values). Right: the Chorus of Fragments (CoF) is a bank of
such fragment modules, each tuned to a particular shape category, appearing in a particular
location [13, 4].
4 Discussion
Human subjects have been previously shown to be able to acquire, through unsupervised
learning, sensitivity to transition probabilities between syllables of nonsense words [7] and
between digits [14], and to co-occurrence statistics of simple geometrical figures [9]. Our
results demonstrate that subjects can also learn (presumably without awareness; cf. [14]) to
treat combinations of complex visual patterns differentially, depending on the conditional
probabilities of the various combinations, accumulated during a short unsupervised training
session.
In our first experiment, the criterion of suspicious coincidence between the occurrences
and
conditions: in each case, we
of and was met in both
had
. Yet, the subjects? behavior indicated a significant
holistic bias: the representation they form tends to be monolithic ( ), unless imperfect
mutual predictability of the potential fragments ( and ) provides support for representing them separately. We note that a similar holistic bias, operating in a setting where a
single encounter with a stimulus can make a difference, is found in language acquisition:
an infant faced with an unfamiliar word will assume it refers to the entire shape of the most
salient object [15]. In our second experiment, both the conditional probabilities as such,
and the suspicious coincidence ratio were found to have the predicted effects, yet these
two factors interacted in a complicated manner, which requires a further investigation.
Our current research focuses on (1) the elucidation of the manner in which subjects process
statistically structured data, (2) the development of the model of structure learning outlined
in the preceding section, and (3) an exploration of the implications of this body of work for
wider issues in vision, such as the computational phenomenology of scene perception [16].
References
[1] H. B. Barlow. Unsupervised learning. Neural Computation, 1:295?311, 1989.
[2] R. S. Zemel and G. E. Hinton. Developing population codes by minimizing description length. Neural Computation, 7:549?564, 1995.
[3] E. Bienenstock, S. Geman, and D. Potter. Compositionality, MDL priors, and object
recognition. In M. C. Mozer, M. I. Jordan, and T. Petsche, editors, Neural Information
Processing Systems, volume 9. MIT Press, 1997.
[4] S. Edelman and N. Intrator. A productive, systematic framework for the representation
of visual structure. In T. K. Leen, T. G. Dietterich, and V. Tresp, editors, Advances in
Neural Information Processing Systems 13, pages 10?16. MIT Press, 2001.
[5] S. E. Palmer. Hierarchical structure in perceptual representation. Cognitive Psychology, 9:441?474, 1977.
[6] M. J. Flannagan, L. S. Fried, and K. J. Holyoak. Distributional expectations and
the induction of category structure. Journal of Experimental Psychology: Learning,
Memory and Cognition, 12:241?256, 1986.
[7] J. R. Saffran, R. N. Aslin, and E. L. Newport. Statistical learning by 8-month-old
infants. Science, 274:1926?1928, 1996.
[8] N. Z. Kirkham, J. A. Slemmer, and S. P. Johnson. Visual statistical learning in infancy:
Evidence for a domain general learning mechanism. Cognition, -:?, 2002. in press.
[9] J. Fiser and R. N. Aslin. Unsupervised statistical learning of higher-order spatial
structures from visual scenes. Psychological Science, 6:499?504, 2001.
[10] SAS. User?s Guide, Version 8. SAS Institute Inc., Cary, NC, 1999.
[11] D. Pomerleau. Input reconstruction reliability estimation. In C. L. Giles, S. J. Hanson, and J. D. Cowan, editors, Advances in Neural Information Processing Systems,
volume 5, pages 279?286. Morgan Kaufmann Publishers, 1993.
[12] I. Stainvas and N. Intrator. Blurred face recognition via a hybrid network architecture.
In Proc. ICPR, volume 2, pages 809?812, 2000.
[13] S. Edelman and N. Intrator. (Coarse Coding of Shape Fragments) + (Retinotopy)
Representation of Structure. Spatial Vision, 13:255?264, 2000.
[14] G. S. Berns, J. D. Cohen, and M. A. Mintun. Brain regions responsive to novelty in
the absence of awareness. Science, 276:1272?1276, 1997.
[15] B. Landau, L. B. Smith, and S. Jones. The importance of shape in early lexical
learning. Cognitive Development, 3:299?321, 1988.
[16] S. Edelman. Constraints on the nature of the neural representation of the visual world.
Trends in Cognitive Sciences, 6:?, 2002. in press.
| 2119 |@word trial:11 version:2 holyoak:1 simulation:2 attended:1 solid:1 accommodate:1 configuration:1 fragment:24 tuned:1 subjective:1 current:2 yet:2 shape:12 infant:4 half:1 fried:1 smith:1 hinged:1 short:1 provides:2 coarse:1 location:3 successive:2 five:2 differential:1 persistent:1 edelman:4 suspicious:12 consists:2 manner:3 mask:6 expected:2 indeed:1 behavior:2 brain:3 landau:1 unfolded:1 considering:1 notation:1 kind:1 lone:1 unified:1 finding:2 sharpening:1 act:1 shed:1 control:1 appear:2 arguably:1 before:4 positive:2 monolithic:3 treat:1 tends:1 tallying:1 twice:1 studied:1 co:5 palmer:2 statistically:2 practice:1 differs:1 digit:1 procedure:3 wholly:1 composite:26 word:5 pre:1 quadrant:1 refers:1 suggest:3 convenience:1 cannot:1 mea:1 context:1 writing:1 lexical:1 exposure:4 attention:1 independently:1 convex:1 assigns:1 insight:1 population:1 notion:2 updated:1 target:15 user:1 us:1 hypothesis:1 associate:1 trend:1 recognition:4 geman:1 labeled:1 distributional:1 bottom:1 module:10 coincidence:12 region:1 principled:1 benjamin:1 predictable:3 mozer:1 cooccurrence:1 asked:1 productive:1 trained:2 depend:1 tight:1 exposed:3 predictive:3 upon:1 basis:1 strikingly:1 joint:2 various:7 represented:4 distinct:2 detected:2 zemel:1 whose:1 larger:2 valued:1 otherwise:1 favor:1 statistic:4 ability:2 sequence:5 blob:1 advantage:1 reconstruction:7 ment:1 peed:5 product:1 interaction:5 date:1 holistic:3 description:2 intervening:1 differentially:1 constituent:8 interacted:1 produce:1 perfect:1 object:16 wider:1 depending:1 measured:4 sa:5 predicted:1 come:1 indicate:2 met:1 concentrate:1 annotated:1 correct:1 exploration:2 human:12 newport:1 disproportionately:2 exchange:1 premise:2 really:1 investigation:2 considered:1 ground:1 credit:2 presumably:1 scope:1 cognition:2 early:1 estimation:2 proc:1 currently:1 propensity:1 sensitive:2 saw:1 vice:2 cary:1 mit:2 cornell:2 sion:1 varying:1 monolithically:1 derived:2 focus:2 consistently:1 indicates:1 attains:1 baseline:9 sense:2 dependent:1 accumulated:1 typically:1 entire:1 bienenstock:1 pixel:2 issue:1 among:2 flexible:1 development:3 spatial:6 mutual:2 marginal:2 field:5 once:1 exerted:1 construct:1 jones:1 unsupervised:14 nearly:1 others:2 stimulus:14 aslin:3 eighty:1 develops:1 composed:1 simultaneously:1 manipulated:1 individual:2 familiar:3 phase:10 consisting:2 maintain:1 ab:2 investigate:1 certainly:1 mdl:8 light:1 behind:1 held:1 implication:1 conforms:1 unless:1 euclidean:1 old:1 psychological:1 earlier:1 giles:1 conducted:1 johnson:1 providence:1 answer:2 combined:1 sensitivity:2 recur:3 probabilistic:2 systematic:1 together:4 reflect:2 cognitive:3 potential:1 coding:2 summarized:1 inc:1 blurred:1 postulated:1 notable:1 explicitly:2 stream:1 script:2 competitive:1 bayes:1 option:2 complicated:3 narrowed:1 parallel:1 possi:1 square:1 variance:3 who:1 kaufmann:1 ensemble:1 judgment:1 yes:1 none:1 served:2 saffran:2 stroke:1 strongest:1 tended:1 acquisition:2 frequency:1 involved:3 con:1 auditory:2 proved:1 pilot:2 knowledge:1 actually:2 appears:2 exceeded:1 higher:7 response:6 maximally:1 leen:1 box:1 just:2 fiser:2 working:1 ally:1 expressive:1 indicated:1 effect:12 dietterich:1 brown:2 verify:1 barlow:7 consisted:2 former:1 hypothesized:1 spatially:1 deal:1 white:1 during:2 criterion:8 m:5 demonstrate:1 geometrical:2 image:13 consideration:1 novel:1 recently:1 functional:1 fourteen:1 cohen:1 volume:3 discussed:1 association:5 analog:3 marginals:1 refer:1 composition:1 measurement:1 versa:2 unfamiliar:3 significant:4 trivially:1 outlined:1 session:3 language:1 had:5 reliability:1 stable:1 access:1 operating:1 own:1 recent:1 showed:2 seen:1 minimum:3 additional:1 morgan:1 preceding:1 determine:1 novelty:1 clockwise:1 signal:1 full:1 faster:2 adapt:1 coded:1 paired:3 controlled:5 watched:1 prediction:2 vision:3 expectation:1 represent:3 sometimes:1 whereas:3 chorus:1 separately:2 participated:2 publisher:1 ithaca:1 sure:1 probably:1 subject:29 tend:1 counterclockwise:1 cowan:1 jordan:1 call:2 structural:1 vw:1 yang:1 presence:1 latin:1 rested:1 revealed:2 independence:4 isolation:1 psychology:3 architecture:2 perfectly:4 imperfect:1 whether:5 generally:2 useful:1 ten:1 category:4 unpaired:2 judging:1 estimated:1 per:1 shall:1 four:3 salient:1 undertaken:1 merely:2 letter:1 throughout:1 reasonable:1 decide:2 decision:1 ble:1 followed:3 distinguish:1 syllable:3 constraint:2 vised:1 ri:1 scene:3 erri:1 dence:1 nathan:2 speedup:9 department:1 structured:3 according:2 developing:1 icpr:1 combination:6 smaller:1 character:15 previously:4 turn:2 mechanism:1 fed:1 serf:1 available:1 phenomenology:1 probe:28 hierarchical:1 intrator:5 petsche:1 occurrence:6 appearing:1 responsive:1 encounter:1 top:1 ensure:1 cf:1 elucidation:1 embodies:1 chinese:6 psychophysical:4 objective:1 quantity:1 looked:1 strategy:1 receptive:5 rt:1 dependence:1 unclear:1 distance:1 separate:1 simulated:2 extent:2 induction:1 potter:1 length:2 code:1 illustration:1 ratio:10 multiplicatively:1 acquire:1 minimizing:1 nc:1 potentially:1 sures:1 negative:1 pomerleau:1 allowing:1 upper:1 observation:1 discarded:1 hinton:1 nonuniform:1 compositionality:1 pair:5 required:2 hanson:1 diction:1 boost:1 adult:1 beyond:1 able:1 alongside:1 pattern:4 perception:1 appeared:4 rf:1 memory:1 treated:1 hybrid:1 representing:2 nonsense:1 auto:2 tresp:1 text:1 prior:3 geometric:1 faced:1 determining:1 embedded:3 mixed:5 localized:1 awareness:2 degree:2 verification:2 experi:1 principle:5 editor:3 bank:1 course:2 placed:1 supported:1 bern:1 accidentally:1 bias:3 allow:1 guide:1 fledged:1 institute:2 face:1 transition:3 stand:2 world:1 collection:1 made:2 replicated:2 simplified:2 adaptive:1 gestalt:2 reconstructed:1 status:1 logic:1 synergy:1 knew:1 alternatively:1 continuous:1 triplet:1 nature:2 learn:2 associator:1 investigated:1 complex:4 domain:3 did:1 significance:1 main:2 motivation:1 whole:3 repeated:2 complementary:1 body:1 screen:1 fashion:1 strengthened:1 ny:1 predictability:3 candidate:1 infancy:1 perceptual:1 third:1 young:1 shimon:1 theorem:1 familiarity:2 symbol:1 evidence:1 structed:1 modulates:1 importance:1 suited:1 appearance:2 likely:1 visual:14 contained:6 corresponds:1 conditional:9 se37:1 month:2 presentation:1 towards:1 absence:1 retinotopy:1 determined:1 specifically:1 unsuper:1 experimental:3 exception:1 indicating:1 internal:2 support:1 latter:3 modulated:2 cof:1 relevance:5 incorporate:1 tested:1 |
1,229 | 212 | 710
Pineda
Time DependentAdaptive Neural Networks
Fernando J. Pineda
Center for Microelectronics Technology
Jet Propulsion Laboratory
California Institute of Technology
Pasadena, CA 91109
ABSTRACT
A comparison of algorithms that minimize error functions to train the
trajectories of recurrent networks, reveals how complexity is traded off for
causality. These algorithms are also related to time-independent
fonnalisms. It is suggested that causal and scalable algorithms are
possible when the activation dynamics of adaptive neurons is fast
compared to the behavior to be learned. Standard continuous-time
recurrent backpropagation is used in an example.
1 INTRODUCTION
Training the time dependent behavior of a neural network model involves the minimization
of a function that measures the difference between an actual trajectory and a desired
trajectory. The standard method of accomplishing this minimization is to calculate the
gradient of an error function with respect to the weights of the system and then to use the
gradient in a minimization algorithm (e.g. gradient descent or conjugate gradient).
Techniques for evaluating gradients and performing minimizations are well developed in the
field of optimal control and system identification, but are only now being introduced to the
neural network community. Not all algorithms that are useful or efficient in control problems
are realizable as physical neural networks. In particular, physical neural network algorithms
must satisfy locality, scaling and causality constraints. Locality simply is the constraint that
one should be able to update each connection using only presynaptic and postsynaptic
infonnation. There should be no need to use infonnation from neurons or connections that
are not in physical contact with a given connection. Scaling, for this paper, refers to the
Time Dependent Adaptive Neural Networks
scaling law that governs the amount of computation or hardware that is required to perform
the weight updates. For neural networks, where the number of weights can become very
large, the amount of hardware or computation required to calculate the gradient must scale
linearly with the number of weights. Otherwise, large networks are not possible. Finally,
learning algorithms must be causal since physical neural networks must evolve forwards in
time. Many algorithms for learning time-dependent behavior, although they are seductively
elegant and computationally efficient, cannot be implemented as physical systems because
the gradient evaluation requires time evolution in two directions. In this paper networks that
violate the causality constraint will be referred to as unphysical.
It is useful to understand how scalability and causality trade off in various gradient evaluation
algorithms. In the next section three related gradient evaluation algorithms are derived and
their scaling and causality properties are compared. The three algorithms demonstrate a
natural progression from a causal algorithm that scales poorly to an a causal algorithm that
scales linearly.
The difficulties that these exact algorithms exhibit appear to be inescapable. This suggests
that approximation schemes that do not calculate exact gradients or that exploit special
properties of the tasks to-be-Ieamed may lead to physically realizable neural networks. The
final section of this paper suggests an approach that could be exploited in systems where the
time scale of the to-be-Ieamed task is much slower than the relaxation time scale of the
adaptive neurons.
2 ANALYSIS OF ALGORITHMS
We will begin by reviewing the learning algorithms that apply to time-dependent recurrent
networks. The control literature generally derives these algorithms by taking a variational
approach (e.g. Bryson and Ho, 1975). Here we will take a somewhat unconventional
approach and restrict oursel yes to the domain of differential equations and their solutions. To
begin with, let us take a concrete example. Consider the neural system given by the equation
,
dx?
(it
=X i+
n
~w
I(x) + I
,=1
I
(1)
Where f(.) is a sigmoid shaped function (e.g. tanh(.)) and ~is an external input This system
is a well studied neural model (e.g. Aplevich, 1968; Cowan, 1967; Hopfield, 1984; Malsburg,
1973; Sejnowski, 1977). The goal is to find the weight matrix w that causes the states x(t)
of the output units to follow a specified trajectory x(t). The actually trajectory depends not
only on the weight matrix but also on the external input vector I. To find the weights one
minimizes a measure of the difference between the actual trajectory x(t) and the desired
trajectory ~(t). This measure is a functional of the trajectories and a function of the weights.
It is given by
tI
1
2
E(w ,t I,t ) =2
dt
,{t) - ~,{t))
(2)
f
.L
,e 0
(x
t
o
where 0 is the set of output units. We shall, only for the purpose of algorithm comparison,
711
712
Pineda
make the following assumptions: (1) That the networks are fully connected (2) That all the
interval [tD,tr] is divided into q segments with numerical integrations performed using the
Euler method and (3) That all the operations are performed with the same precision. This will
allow us to easily estimate the amount of computation and memory required for each
algorithm relative to the others.
2.1 ALGORITHM A
If the objective function E is differentiated with respect to wn one obtains
-
aE
aw
=rs
Lnft!d t J i(t) P irit )
t
i=1
where
and where
o
gi(t)- x i(t)
J.= {
(3a)
'0
if i E 0
ififl.O
ax ,?
Pirs=-a-
(3b)
(3c)
Wrs
To evaluate Pirs' differentiate equation (1) with respect to wn and observe that the time
derivative and the partial derivative with respect to wn commute. The resulting equation is
dp irs ~L ( )
- d = ~ ij'X j Pjrs+Sir.
where
t
.1
(4a)
J=
(4b)
and where
(4c)
=
The initial condition for eqn. (4a) is p(t) O. Equations (1), (3) and (4) can be used to
calculate the gradient for a learning rule. This is the approach taken by Williams and Zipser
(1989) and also discussed by Pearlmutter(1988). Williams and Zipser further observe that
one can use the instantaneous value of p(t) and J(t) to update the weights continually
provided the weights change slowly. The computationally intensive part of this algorithm
occurs in the integration of equation (4a). There are n3 components to p hence there are Ji3
equations . Accordingly the amount of hardware or memory required to perform the
calculation will scale like n 3? Each of these equations requires a summation over all the
neurons, hence the amount of computation (measured in multiply-accumulates) goes like It
per time step, and there are q time steps, hence the total number of multiply-accumulates
scales like n4q Clearly, the scaling properties of this approach are very poor and it cannot
be practically applied to very large networks.
2.2 ALGORITHM B
Rather than numerically integrate the system of equations (4a) to obtain p(t), suppose we
write down the formal solution. This solution is
Time Dependent Adaptive Neural Networks
11
Pirs(t)='LKij(t,to)PjrsCt
0)+
j=1
'L"f'drKjj(t,f)Sjrs(i)
j=1
'0
(Sa)
The matrix K is defined by the expression
p(.r.. '~T
,) = ex
K (' 2'
L (x (T)))
(5b)
This matrix is known as the propagator or transition matrix. The expression for Pit. consists
of a homogeneous solution and a particular solution. The choice of initial condition Pirs(to)
0 leaves only the particular solution. If the particular solution is substituted back into eqn.
(3a), one eventually obtains the following expression for the gradient
=
aE
11
- = - 'Lf
'f
f '
d-r J;Ct)K irU ,-r)f(x s(-r))
(6)
rs
j=1
'0
'0
To obtain this expression one must observe that s.In can be expressed in terms of x? , i.e. use
eqn. (4c). This allows the summation over j to be performed trivially, thus resulting in
eqn.(6). The familiar outer product form of backpropagation is not yet manifest in this
expression. To uncover it, change the order of the integrations. This requires some care
because the limits of the integration are not the same. The result is
aw
aE
'L
11
-=-
aw
dt
f
If
d-rf
If
dt Jj(t)K irU ,-r)f(x sC-r))
(7)
rs
i=1
'0
l'
Inspection ofthis expression reveals that neither the summation over i nor the integration over
't includes x.(t), thus it is useful to factor it out. Consequently equation (7) takes on the
familiar outer product form of backpropagation
aE
If
-= -f
aw
rs
dt Y r(t)f(x sU))
(8)
l'
Where yr(t) is defined to be
If
11
Y r(-r) =-
'L f
i= 1
dt Jj(t)K irU ,-r)
(9)
t'
Equation (8), defines an expression for the gradient, provided we can calculate Yr(t) from eqn.
(9). In principle, this can be done since the propagator K and the vector J are both completely
determined by x(t). The computationally intensive part of this algorithm is the calculation
of K(t, 't) for all values of t and't. The calculation requires the integration of equations of the
form
(10)
dK ,-r) - L (x U) K (t ,-r)
i:
for q different values of't. There are n2different equations to integrate for each value of't
Consequently there are n2q integrations to be performed where the interval from to to tf is
divided into q intervals. The calculation of all the components ofK(t,'t), from tr to t ,scales
like n3q2, since each integration requires n multiply-accumulates per time step and there are
q time steps. Similarly, the memory requirements scale like n2q2. This is because K has n2
components for each (t,'t) pair and there are q2 such pairs.
713
714
Pineda
Equation (10) must be integrated forwards in time from t= 't to t = trand backwards in time
from t= 't to t = to. This is because K must satisfy K( 't?'t) = 1 (the identity matrix) for all
'to This condition follows from the definition of K eqn. (5b). Finally, we observe that
expression (9) is the time-dependent analog of the expression used by Rohwer and Forrest
(1987) to calculate the gradient in recurrent networks. The analogy can be made somewhat
more explicit by writingK(t,'t) as the inverse K-l('t,t). Thus we see that y( t) can be expressed
in terms of a matrix inverse just as in the Rohwer and Forrest algorithm.
2.3
ALGORITHM C
The final algorithm is familiar from continuous time optimal control and identification. The
algorithm is usually derived by performing a variation on the functional given by eqn. (2).
This results in a two-point boundary value problem. On the other hand, we know that y is
given by eqn. (9). So we simply observe that this is the particular solution of the differential
equation
dy
T
- ([t= L (x (t))y +J
(11)
Where LT is the transpose of the matrix defined in eqn. (4b). To see this simply substitute
the form for y into eqn. (11) and verify that it is indeed the solution to the equation.
The particular solution to eqn. (11) vanishes only if y(1r) = O. In other words: to obtain yet)
we need only integrate eqn. (11) backwards from the final condition y(t~ = O. This is just
the algorithm introduced to the neural network community by Pearlmutter (1988). This also
corresponds to the unfolding in time approach discussed by Rumelhart et al. (1986), provided
that all the equations are discretized and one takes At = 1.
The two point boundary value problem is rather straight forward to solve because the
equation for x(t) is independent of yet). Both x(t) and yet) can be obtained with n multiplyaccumulates per time step. There are q time steps from to to tfand bothx(t) and yet) have n
components, hence the calculation of x(t) and yet) scales like 02q. The weight update
equation also requires n2q mUltiply- accumulates. Thus the computational requirements of
the algorithm as a whole scale like n2q The memory required also scales like n2q, since it
is necessary to save each value of x(t) along the trajectory to compute yet).
2.4
SCALING VS CAUSALITY
The results of the previous sections are summarized in table 1 below. We see that we have
a progression of tradeoffs between scaling and causality. That is, we must choose between
a causal algorithm with exploding computational and storage requirements and an a causal
algorithm with modest storage requirements. There is no q dependence in the memory
requirments because the integral given in eqn. (3a) can be accumulated at each time step.
Algorithm B has some of the worst features of both algorithms.
Time Dependent Adaptive Neural Networks
Table 1: Comparison of three algorithms
Algorithm
A
B
C
Memory
Multiply
-accumulates
diirection of integations
x and p are both forward in time
x is forward, K is forward and backward
x is forward, y is backward in time.
Digital hardware has no difficulties (at least over finite time intervals) with a causal
algorithms provided a stack is available to act as a memory that can recall states in reverse
order. To the extent that the gradient calculations are carried out on digital machines, it makes
sense to use algorithm C because it is the most efficient. In analog VLSI however, it is
difficult to imagine how to build a continually running network that uses an a causal
algorithm. Algorithm A is attractive for physical implementation because it could be run
continually and in real time (Williams and Zipser, 1989). However, its scaling properties
preclude the possibility of building very large networks based on the algorithm. Recently,
Zipser (1990) has suggested that a divide and conquer approach may reduce the
computational and spatial complexity of the algorithm. This approach, although promising,
does not always work and there is as yet no convergence proof. How then, is it possible to
learn trajectories using local, scalable and causal algorithms? In the next section a possible
avenue of attack is suggested.
3 EXPLOITING DISPARATE TIME SCALES
I assert that for some classes of problems there are scalable and causal algorithms that
approximate the gradient and that these algorithms can be found by exploiting the disparity
in time scales found in these classes of problems. In particular, I assert that when the time
scale of the adaptive units is fast compared to the time scale of the behavior to be learned, it
is possible to find scalable and causal adaptive algorithms. A general formalism for doing
this will not be presented here, instead a simple, perhaps artificial, example will be presented.
This example minimizes an error function for a time dependent problem.
It is likely that trajectory generation in motor control problems are of this type. The
characteristic time scales of the trajectories that need to be generated are determined by
inertia and friction. These mechanical time scales are considerably longer than the electronic
time scales that occur in VLSI. Thus it seems that for robotic problems, there may be no need
to use the completely general algorithms discussed in section 2. Instead, algorithms that take
advantage of the disparity between the mechanical and the electronic time scales are likely
to be more useful for learning to generate trajectories.
he task is to map from a periodic input I(t) to a periodic output ~(t). The basic idea is to use
the continuous-time recurrent-backpropagation approach with slowly varying timedependent inputs rather than with static inputs. The learning is done in real-time and in a
continuous fashion. Consider a set of n "fast" neurons (i= 1,.. ,n) each of which satisfies the
715
716
Pineda
additive activation dynamics determined by eqn (1). Assume that the initial weights are
sufficientl y small that the dynamics of the network would be convergent if the inputs I were
constant. The external input vector ~ is applied to the network through the vector I. It has
been previously shown (pineda, 1988) that the ij-th component of the gradient ofE is equal
to yfjf(xf) where Xfj is the steady state solution of eqn. (1) and where yfjis a component of
the steady state solution of
dy
T
f
(12)
- = L (x )y +1
dt
where the components ofLT are given by eqn. (4.b). Note that the relative sign between
equations (11) and (12) is what enables this algorithm to be causal. Now suppose that instead
of a fixed input vector I, we use a slowly varying input I(t/'t ) where't is the characteristic
time scale over which the input changes significantly. If w~ take as lite gradient descent
algorithm, the dynamics defined by
dw
rs
't'w([t=Y i(t)X /t)
(13)
where't.. is the time constant that defines the (slow) time scale over which w changes and
where Xj is the instantaneous solution of eqn. (1) and Yj is the instantaneous solution of
eqn.(12) . Then in the adiabatic limit the Cartesian product yl(x) in eqn. (13) approximates
the negative gradient of the objective function E, that is
(14)
This approach can map one continuous trajectory into another continuous trajectory,
provided the trajectories change slowly enough. Furthermore, learning occurs causally and
scalably. There is no memory in the model, i.e. the output of the adaptive neurons depends
only on their input and not on their internal state. Thus, this network can never learn to
perform tasks that require memory unless the learning algorithm is modified to learn the
appropriate transitions. This is the major drawback of the adiabatic approach. Some state
information can be incorporated into this model by using recurrent connections - in which
case the network can have multiple basins and the final state will depend on the initial state
of the net as well as on the inputs, but this will not be pursued here.
Simple simulations were performed to verify that the approach did indeed perform gradient
descent. One simulation is presented here for the benefit of investigators who may wish to
verify the results. A feedforward network topology consisting of two input units, five hidden
units and two output units was used for the adaptive network. Units were numbered
sequentially, 1 through 9, beginning with the input layer and ending in the output layer. Time
dependent external inputs for the two input neurons were generated with time dependence
II = sin(27tt) and ~ = cos(2m). The targets for the output neurons were ~ = R sin(27tt) and
~9 =R cos(2m) where R = 1.0 + 0.lsin(6m). All the equations were simultaneously integrated
using 4th order Runge-Kutta with a time step of 0.1. A relaxation time scale was introduced
into the forward and backward propagation equations by multiplying the time derivatives in
eqns. (1) and (12) by't" and 'tyrespectively. These time scales were set to't" ='ty= 0.5. The
adaptive time scale of the weights was 't.. = 1.0. The error in the network was initially, E =
Time Dependent Adaptive Neural Networks
10 and the integration was cut off when the error reached a plateau at E = 0.12. The learning
curve is shown in Fig. 1. The trained trajectory did not exactly reach the desired solution. In
particular the network did not learn the odd order hannonic that modulates R. By way of
comparison, a conventional backpropagation approach that calculated a cumulative gradient
over the trajectory and used conjugate gradient for the descent, was able to converge to the
global minimum.
12,---------------------------------~
10-'
III
III
8 6 4-
m
m
m
m
m
m
ED
m
2 -
.
O+-__~---~~??~??E??~?.B..~..B..B. .~?.m??D?~???~~
o
I
I
10
20
I
30
40
50
Time
Figure 1: Learning curve. One time unit corresponds to a single oscillation
4 SUMMARY
The key points of this paper are: 1) Exact minimization algorithms for learning timedependent behavior either scale poorl y or else violate causality and 2) Approximate gradient
calculations will likely lead to causal and scalable learning algorithms. The adiabatic
approach should be useful for learning to generate trajectories of the kind encountered when
learning motor skills.
References herein to any specific commercial product, process, or service by trade name, trademark, manufacturer,
or otherwise, does not constitute or imply any endorsement by the Uoited States Government or the Jet Propulsion
Laboratory, California Institute of Technology. The work described in this paper was carried out at the
Center for Space Microelectonrics Technology, Jet Propulsion Laboratory, California Institute of
Technology. Support for the work came from the Air Force Office of Scientific Research through an
agreement with the National Aeronautics and Space Administration (AFOSR-ISSA-90-0027).
REFERENCES
Aplevich,J.D. (1968). Models of certain nonlinear systems. InE.R.Caianiello(Ed.),Neural
Networks, (pp. 110-115), Berlin: Springer Verlag.
Bryson, A. E. and Ho, Y. (1975). Applied Optimal Control: Optimization. Estimation. and
717
718
Pineda
Control. New York: Hemisphere Publishing Co.
Cowan, J. D. (1967). A mathematical theory of central nervous activity. Unpublished
dissertation, Imperial College, University of London.
Hopfield, J. J. (1984). Neurons with graded response have collective computational
properties like those of two-state neurons. Proc. Nat. Acad. Sci. USA, Bio., ..8.l. 3088-3092.
Malsburg, C. van der (1973). Self-organization of orientation sensitive cells in striate cortex,
Kybernetic, 14,85-100.
Pearlmutter, B. A. (1988), Learning state space trajectories in recurrent neural networks: A
preliminary report, (Tech. Rep. AlP-54), Department of Computer Science , Carnegie Mellon
University, Pittsburgh, PA
Pineda, F. J. (1988). Dynamics and Architecture for Neural Computation. Journal of
Complexity,~, (pp.216-245)
Rowher R, R. and Forrest, B. (1987). Training time dependence in neural networks, In M.
CaudilandC.Butler,(Eds.),ProceedingsoftheIEEEFirstAnnuallnternationalConference
on Neural Networks, ~, (pp. 701-708). San Diego, California: IEEE.
Rumelhart, D. E., Hinton, G. E., and Willaims, R.J. (1986). Learning Internal
Representations by Error Propagation. In D. E. Rumelhart and J. L. McClelland, (Eds.),
Parallel Distributed Processing, (pp. 318-362). Cambridge: M.LT. Press.
Sejnowski, T. J. (1977). Storing covariance with nonlinearly interacting neurons. Journal
of Mathematical Biology, ~,303 .. 321.
Williams, R.I. and Zipser, D. (1989). A learning algorithm for continually running
fully recurrent neural networks. Neural Computation, 1, (pp. 270-280).
Zipser, D. (1990). Subgrouping reduces complexity and speeds up learning in recurrent
networks, (this volume).
| 212 |@word seems:1 scalably:1 simulation:2 r:5 covariance:1 hannonic:1 commute:1 tr:2 initial:4 disparity:2 activation:2 yet:8 dx:1 must:8 additive:1 numerical:1 enables:1 motor:2 update:4 v:1 pursued:1 leaf:1 yr:2 nervous:1 accordingly:1 inspection:1 beginning:1 dissertation:1 attack:1 five:1 mathematical:2 along:1 become:1 differential:2 consists:1 indeed:2 behavior:5 nor:1 discretized:1 td:1 actual:2 preclude:1 begin:2 provided:5 what:1 kind:1 minimizes:2 q2:1 developed:1 assert:2 ti:1 act:1 exactly:1 control:7 unit:8 bio:1 appear:1 causally:1 continually:4 service:1 local:1 limit:2 acad:1 accumulates:5 studied:1 suggests:2 pit:1 co:3 yj:1 timedependent:2 lf:1 backpropagation:5 ofe:1 significantly:1 word:1 refers:1 numbered:1 cannot:2 storage:2 conventional:1 map:2 center:2 williams:4 go:1 rule:1 dw:1 variation:1 imagine:1 suppose:2 target:1 commercial:1 exact:3 diego:1 homogeneous:1 us:1 agreement:1 pa:1 rumelhart:3 cut:1 worst:1 calculate:6 connected:1 trade:2 vanishes:1 complexity:4 dynamic:5 caianiello:1 trained:1 depend:1 reviewing:1 segment:1 completely:2 easily:1 hopfield:2 various:1 train:1 fast:3 london:1 sejnowski:2 artificial:1 sc:1 solve:1 otherwise:2 gi:1 final:4 runge:1 pineda:8 differentiate:1 advantage:1 net:1 product:4 poorly:1 ine:1 scalability:1 exploiting:2 convergence:1 requirement:4 recurrent:9 measured:1 ij:2 odd:1 ji3:1 sa:1 implemented:1 involves:1 direction:1 drawback:1 alp:1 require:1 government:1 preliminary:1 summation:3 practically:1 traded:1 major:1 purpose:1 estimation:1 proc:1 tanh:1 infonnation:2 sensitive:1 tf:1 minimization:5 unfolding:1 clearly:1 always:1 modified:1 rather:3 varying:2 office:1 derived:2 ax:1 tech:1 bryson:2 realizable:2 sense:1 dependent:10 accumulated:1 integrated:2 initially:1 pasadena:1 vlsi:2 hidden:1 orientation:1 spatial:1 special:1 integration:9 field:1 equal:1 never:1 shaped:1 n2q:4 biology:1 others:1 report:1 simultaneously:1 national:1 lite:1 familiar:3 consisting:1 irs:1 organization:1 possibility:1 multiply:5 evaluation:3 subgrouping:1 integral:1 partial:1 necessary:1 modest:1 unless:1 divide:1 desired:3 causal:13 formalism:1 euler:1 aw:4 periodic:2 considerably:1 off:3 yl:1 concrete:1 central:1 choose:1 slowly:4 external:4 derivative:3 issa:1 summarized:1 includes:1 satisfy:2 depends:2 performed:5 doing:1 reached:1 wrs:1 parallel:1 minimize:1 air:1 accomplishing:1 characteristic:2 who:1 yes:1 identification:2 trajectory:20 multiplying:1 straight:1 plateau:1 reach:1 ed:4 definition:1 rohwer:2 ty:1 pp:5 proof:1 static:1 manifest:1 recall:1 uncover:1 actually:1 back:1 dt:6 follow:1 response:1 done:2 furthermore:1 just:2 hand:1 eqn:19 su:1 nonlinear:1 inescapable:1 propagation:2 defines:2 perhaps:1 scientific:1 name:1 usa:1 building:1 verify:3 evolution:1 hence:4 laboratory:3 attractive:1 sin:2 self:1 eqns:1 steady:2 tt:2 demonstrate:1 pearlmutter:3 variational:1 instantaneous:3 recently:1 sigmoid:1 functional:2 physical:6 volume:1 discussed:3 analog:2 approximates:1 he:1 numerically:1 mellon:1 cambridge:1 trivially:1 similarly:1 longer:1 cortex:1 aeronautics:1 hemisphere:1 reverse:1 certain:1 verlag:1 rep:1 came:1 der:1 exploited:1 minimum:1 somewhat:2 care:1 converge:1 fernando:1 exploding:1 ii:1 violate:2 multiple:1 reduces:1 xf:1 jet:3 calculation:7 divided:2 scalable:5 basic:1 ae:4 physically:1 cell:1 interval:4 else:1 irit:1 lsin:1 elegant:1 cowan:2 zipser:6 backwards:2 feedforward:1 iii:2 enough:1 wn:3 xj:1 architecture:1 restrict:1 topology:1 reduce:1 idea:1 avenue:1 tradeoff:1 intensive:2 administration:1 expression:9 york:1 cause:1 jj:2 constitute:1 useful:5 generally:1 governs:1 amount:5 hardware:4 mcclelland:1 generate:2 sign:1 per:3 write:1 carnegie:1 shall:1 key:1 imperial:1 neither:1 backward:3 relaxation:2 run:1 inverse:2 electronic:2 forrest:3 oscillation:1 endorsement:1 dy:2 scaling:8 layer:2 ct:1 convergent:1 encountered:1 activity:1 occur:1 constraint:3 n3:1 speed:1 friction:1 performing:2 department:1 poor:1 conjugate:2 postsynaptic:1 taken:1 computationally:3 equation:22 previously:1 eventually:1 know:1 unconventional:1 available:1 operation:1 apply:1 progression:2 observe:5 manufacturer:1 differentiated:1 appropriate:1 save:1 ho:2 slower:1 substitute:1 ofk:1 running:2 publishing:1 malsburg:2 exploit:1 fonnalisms:1 build:1 conquer:1 graded:1 contact:1 objective:2 occurs:2 dependence:3 striate:1 exhibit:1 gradient:23 dp:1 kutta:1 berlin:1 sci:1 outer:2 propulsion:3 presynaptic:1 extent:1 difficult:1 negative:1 disparate:1 implementation:1 collective:1 perform:4 neuron:11 finite:1 descent:4 propagator:2 hinton:1 incorporated:1 interacting:1 stack:1 community:2 introduced:3 pair:2 required:5 specified:1 mechanical:2 connection:4 unpublished:1 nonlinearly:1 california:4 learned:2 herein:1 able:2 suggested:3 usually:1 below:1 rf:1 memory:9 natural:1 difficulty:2 force:1 scheme:1 technology:5 imply:1 carried:2 literature:1 evolve:1 relative:2 law:1 sir:1 fully:2 afosr:1 generation:1 analogy:1 digital:2 integrate:3 basin:1 principle:1 storing:1 summary:1 transpose:1 formal:1 allow:1 understand:1 institute:3 taking:1 distributed:1 benefit:1 van:1 boundary:2 curve:2 calculated:1 evaluating:1 transition:2 ending:1 cumulative:1 forward:8 made:1 adaptive:11 inertia:1 san:1 approximate:2 obtains:2 skill:1 global:1 robotic:1 reveals:2 sequentially:1 pittsburgh:1 butler:1 continuous:6 table:2 promising:1 learn:4 ca:1 domain:1 substituted:1 did:3 linearly:2 whole:1 n2:1 causality:8 referred:1 fig:1 fashion:1 slow:1 precision:1 adiabatic:3 explicit:1 wish:1 down:1 specific:1 trademark:1 dk:1 microelectronics:1 derives:1 ofthis:1 modulates:1 nat:1 cartesian:1 locality:2 lt:2 simply:3 likely:3 expressed:2 springer:1 corresponds:2 satisfies:1 goal:1 identity:1 consequently:2 change:5 determined:3 unphysical:1 total:1 college:1 internal:2 support:1 investigator:1 evaluate:1 ex:1 |
1,230 | 2,120 | Bayesian Predictive Profiles with
Applications to Retail Transaction Data
Igor V. Cadez
Information and Computer Science
University of California
Irvine, CA 92697-3425, U.S.A.
[email protected]
Padhraic Smyth
Information and Computer Science
University of California
Irvine, CA 92697-3425, U.S.A.
[email protected]
Abstract
Massive transaction data sets are recorded in a routine manner in
telecommunications, retail commerce, and Web site management.
In this paper we address the problem of inferring predictive individual profiles from such historical transaction data. We describe a generative mixture model for count data and use an an
approximate Bayesian estimation framework that effectively combines an individual?s specific history with more general population
patterns. We use a large real-world retail transaction data set to
illustrate how these profiles consistently outperform non-mixture
and non-Bayesian techniques in predicting customer behavior in
out-of-sample data.
1
Introduction
Transaction data sets consist of records of pairs of individuals and events, e.g., items
purchased (market basket data), telephone calls made (call records), or Web pages
visited (from Web logs). Of significant practical interest in many applications is
the ability to derive individual-specific (or personalized) models for each individual from the historical transaction data, e.g., for exploratory analysis, adaptive
personalization, and forecasting.
In this paper we propose a generative model based on mixture models and Bayesian
estimation for learning predictive profiles. The mixture model is used to address
the heterogeneity problem: different individuals purchase combinations of products
on different visits to the store. The Bayesian estimation framework is used to
address the fact that we have different amounts of data for different individuals.
For an individual with very few transactions (e.g., only one) we can ?shrink? our
predictive profile for that individual towards a general population profile. On the
other hand, for an individual with many transactions, their predictive model can be
more individualized. Our goal is an accurate and computationally efficient modeling
framework that smoothly adapts a profile to each individual based on both their own
historical data as well as general population patterns. Due to space limitations only
selected results are presented here; for a complete description of the methodology
and experiments see Cadez et al. (2001).
The idea of using mixture models as a flexible approach for modeling discrete and
categorical data has been known for many years, e.g., in the social sciences for latent
class analysis (Lazarsfeld and Henry, 1968). Traditionally these methods were only
applied to relatively small low-dimensional data sets. More recently there has been
a resurgence of interest in mixtures of multinomials and mixtures of conditionally
independent Bernoulli models for modeling high-dimensional document-term data
in text analysis (e.g., McCallum, 1999; Hoffman, 1999). The work of Heckerman
et al. (2000) on probabilistic model-based collaborative filtering is also similar in
spirit to the approach described in this paper except that we focus on explicitly
extracting individual-level profiles rather than global models (i.e., we have explicit
models for each individual in our framework). Our work can be viewed as being an
extension of this broad family of probabilistic modeling ideas to the specific case
of transaction data, where we deal directly with the problem of making inferences
about specific individuals and handling multiple transactions per individual. Other
approaches have also been proposed in the data mining literature for clustering
and exploratory analysis of transaction data, but typically in a non-probabilistic
framework (e.g., Agrawal, Imielinski, and Swami, 1993; Strehl and Ghosh, 2000;
Lawrence et al., 2001). The lack of a clear probabilistic semantics (e.g., for association rule techniques) can make it difficult for these models to fully leverage the
data for individual-level forecasting.
2
Mixture-Basis Models for Profiles
We have an observed data set D = {D1 , . . . , DN }, where Di is the observed data on
the ith customer, 1 ? i ? N . Each individual data set Di consists of one or more
transactions for that customer , i.e., Di = {yi1 , . . . , yij , . . . , yini }, where yij is the
jth transaction for customer i and ni is the total number of transactions observed
for customer i.
The jth transaction for individual i, yij , consists of a description of the set of
products (or a ?market basket?) that was purchased at a specific time by customer
i (and yi will be used to denote an arbitrary transaction from individual i). For
the purposes of the experiments described in this paper, each individual transaction
yij is represented as a vector of d counts yij = (nij1 , . . . nijc , . . . , nijC ), where nijc
indicates how many items of type c are in transaction yij , 1 ? c ? C.
We define a predictive profile as a probabilistic model p(yi ), i.e., a probability
distribution on the items that individual i will purchase during a store-visit. We
propose a simple generative mixture model for an individual?s purchasing behavior,
namely that a randomly selected transaction yi from individual i is generated by one
of K components in a K-component mixture model. The kth mixture component,
1 ? k ? K is a specific model for generating the counts and we can think of each of
the K models as ?basis functions? describing prototype transactions. For example,
in a clothing store, one might have a mixture component that acts as a prototype
for suit-buying behavior, where the expected counts for items such as suits, ties,
shirts, etc., given this component, would be relatively higher than for the other
items.
There are several modeling choices for the component transaction models for generating item counts. In this paper we choose a particularly simple memoryless
multinomial model that operates as follows. Conditioned on nij (the total number
of items in the basket) each of the individual items is selected in a memoryless
fashion by nij draws from a multinomial distribution Pk = (?k1 , . . . , ?kC ) on the C
possible items, ?kj = 1.
Probability
0.6
0.6
COMPONENT 1
0.4
0
0.2
0
10
20
30
40
50
Probability
0.6
COMPONENT 3
10
20
30
40
50
COMPONENT 4
0.2
0
10
20
30
40
50
0.6
Probability
0
0.4
0.2
0
0
10
20
30
40
50
0.6
COMPONENT 5
0.4
COMPONENT 6
0.4
0.2
0
0
0.6
0.4
0
COMPONENT 2
0.4
0.2
0.2
0
10
20
30
Department
40
50
0
0
10
20
30
Department
40
50
Figure 1: An example of 6 ?basis? mixture components fit to retail transaction
data.
Figure 1 shows an example of K = 6 such basis mixture components that have
been learned from a large retail transaction data (more details on learning will be
discussed below). Each window shows a different set of component probabilities
Pk , each modeling a different type of transaction. The components show a striking
bimodal pattern in that the multinomial models appear to involve departments
that are either above or below department 25, but there is very little probability
mass that crosses over. In fact the models are capturing the fact that departments
numbered lower than 25 correspond to men?s clothing and those above 25 correspond
to women?s clothing, and that baskets tend to be ?tuned? to one set or the other.
2.1
Individual-Specific Weights
We further assume that for each individual i there exists a set of K weights,
and in the general case
P these weights are individual-specific, denoted by ?i =
(?i1 , . . . , ?iK ), where k ?ik = 1. Weight ?ik represents the probability that when
individual i enters the store their transactions will be generated by component k.
Or, in other words, the ?ik ?s govern individual i?s propensity to engage in ?shopping
behavior? k (again, there are numerous possible generalizations, such as making the
?ik ?s have dependence over time, that we will not discuss here). The ?ik ?s are in
effect the profile coefficients for individual i, relative to the K component models.
This idea of individual-specific weights (or profiles) is a key component of our proposed approach. The mixture component models Pk are fixed and shared across
all individuals, providing a mechanism for borrowing of strength across individual
data. The individual weights are in principle allowed to freely vary for each individual within a K-dimensional simplex. In effect the K weights can be thought as
basis coefficients that represent the location of individual i within the space spanned
by the K basis functions (the component Pk multinomials). This approach is quite
similar in spirit to the recent probabilistic PCA work of Hofmann (1999) on mixture
models for text documents, where he proposes a general mixture model framework
that represents documents as existing within a K-dimensional simplex of multinomial component models.
The model for each individual is an individual-specific mixture model, where the
Number of items
8
TRAINING PURCHASES
6
4
2
0
0
5
10
15
20
25
30
35
40
45
50
40
45
50
Number of items
8
TEST PURCHASES
6
4
2
0
0
5
10
15
20
25
Department
30
35
Figure 2: Histograms indicating which products a particular individual purchased,
from both the training data and the test data.
0.2
Probability
PROFILE FROM GLOBAL WEIGHTS
0.15
0.1
0.05
0
0
5
10
15
20
25
30
35
40
45
50
40
45
50
40
45
50
0.2
Probability
SMOOTHED HISTOGRAM PROFILE (MAP)
0.15
0.1
0.05
0
0
5
10
15
20
25
30
35
0.2
Probability
PROFILE FROM INDIVIDUAL WEIGHTS
0.15
0.1
0.05
0
0
5
10
15
20
25
30
35
Figure 3: Inferred ?effective? profiles from global weights, smoothed histograms, and
individual-specific weights for the individual whose data was shown in Figure 2.
weights are specific to individual i:
p(yij )
=
K
X
?ik p(yij |k)
k=1
=
K
X
k=1
?ik
C
Y
n
?kcijc .
c=1
where ?kc is the probability that the cth item is purchased given component k
and nijc is the number of items of category c purchased by individual i, during
transaction ij.
As an example of the application of these ideas, in Figure 2 the training data and
test data for a particular individual are displayed. Note that there is some predictability from training to test data, although the test data contains (for example)
a purchase in department 14 (which was not seen in the training data). Figure 3
plots the effective profiles1 for this particular individual as estimated by three different schemes in our modeling approach: (1) global weights that result in everyone
1
We call these ?effective profiles? since the predictive model under the mixture assump-
being assigned the same ?generic? profile, i.e., ?ik = ?k , (2) a maximum a posteriori (MAP) technique that smooths each individual?s training histogram with
a population-based histogram, and (3) individual weights estimated in a Bayesian
fashion that are ?tuned? to the individual?s specific behavior. (More details on
each of these methods are provided later in the paper; a complete description can
be found in Cadez et al. (2001)).
One can see in Figure 3 that the global weight profile reflects broad population-based
purchasing patterns and is not representative of this individual. The smoothed
histogram is somewhat better, but the smoothing parameter has ?blurred? the
individual?s focus on departments below 25. The individual-weight profile appears
to be a better representation of this individual?s behavior and indeed it does provide
the best predictive score (of the 3 methods) on the test data in Figure 2. Note that
the individual-weights profile in Figure 3 ?borrows strength? from the purchases of
other similar customers, i.e., it allows for small but non-zero probabilities of the
individual making purchases in departments (such as 6 through 9) if he or she has
not purchased there in the past. This particular individual?s weights, the ?ik ?s, are
(0.00, 0.47, 0.38, 0.00, 0.00.0.15) corresponding to the six component models shown
in Figure 1. The most weight is placed on components 2, 3 and 6, which agrees
with our intuition given the individual?s training data.
2.2
Learning the Model Parameters
The unknown parameters in our model consist of both the parameters of the K
multinomials, ?kc , 1 ? k ? K, 1 ? c ? C, and the vectors of individual-specific
profile weights ?i , 1 ? i ? N . We investigate two different approaches to learning
individual-specific weights:
? Mixture-Based Maximum Likelihood (ML) Weights: We treat the
weights ?i and component parameters ? as unknown and use expectationmaximization (EM) to learn both simultaneously. Of course we expect this
model to overfit given the number of parameters being estimated but we
include it nonetheless as a baseline.
? Mixture-Based Empirical Bayes (EB) Weights: We first use EM
to learn a mixture of K transaction models (ignoring individuals). We
then use a second EM algorithm in weight-space to estimate individualspecific weights ?i for each individual. The second EM phase uses a fixed
empirically-determined prior (a Dirichlet) for the weights. In effect, we are
learning how best to represent each individual within the K-dimensional
simplex of basis components. The empirical prior uses the marginal weights
(??s) from the first run for the mean of the Dirichlet, and an equivalent
sample size of n = 10 transactions is used in the results reported in the
paper. In effect, this can be viewed as an approximation to either a fully
Bayesian hierarchical estimation or an empirical Bayesian approach (see
Cadez et al. (2001) for more detailed discussion). We did not pursue the
fully Bayesian or empirical Bayesian approaches for computational reasons
since the necessary integrals cannot be evaluated in closed form for this
model and numerical methods (such as Markov Chain Monte Carlo) would
be impractical given the data sizes involved.
We also compare two other approaches for profiling for comparison: (1) Global
Mixture Weights: instead of individualized weights we set each individual?s
tion is not a multinomial that can be plotted as a bar chart: however, we can approximate
it and we are plotting one such approximation here
3.5
3.4
Negative LogP Score [bits/token]
3.3
Individualized MAP weights
3.2
3.1
Mixtures: Individualized ML weights
3
Mixtures: Global mixture weights
2.9
2.8
2.7
Mixtures: Individualized EB weights
2.6
2.5
0
10
20
30
40
50
60
70
Number of Mixture Components [k]
80
90
100
Figure 4: Plot of the negative log probability scores per item (predictive entropy) on
out-of-sample transactions, for various weight models as a function of the number
of mixture components K.
weight vector to the marginal weights (?i k = ?k ), and (2) Individualized MAP
weights: a non-mixture approach where we use an empirically-determined Dirichlet prior directly on the multinomials, and where the equivalent sample size of this
prior was ?tuned? on the test set to give optimal performance. This provides an (optimistic) baseline of using multinomial profiles directly, without use of any mixture
models.
3
Experimental Results
To evaluate our approach we used a real-world transaction data set. The data
consists of transactions collected at a chain of retail stores over a two-year period.
We analyze the transactions here at the store department level (50 categories of
items). We separate the data into two time periods (all transactions are timestamped), with approximately 70% of the data being in the first time period (the
training data) and the remainder in the test period data. We train our mixture and
weight models on the first period and evaluate our models in terms of their ability
to predict transactions that occur in the subsequent out-of-sample test period.
The training data contains data on 4339 individuals, 58,866 transactions, and
164,000 items purchased. The test data consists of 4040 individuals, 25,292 transactions, and 69,103 items purchased. Not all individuals in the test data set appear
in the training data set (and vice-versa): individuals in the test data set with no
training data are assigned a global population model for scoring purposes.
To evaluate the predictive power of each model, we calculate the log-probability
(?logp scores?) of the transactions as predicted by each model. Higher logp scores
mean that the model assigned higher probability to events that actually occurred.
Note that the mean negative logp score over a set of transactions, divided by the
total number of items, can be interpreted as a predictive entropy term in bits. The
lower this entropy term, the less uncertainty in our predictions (bounded below by
zero of course, corresponding to zero uncertainty).
Figure 4 compares the out-of-sample predictive entropy scores as a function of the
0
?50
?55
?50
?60
?100
logP, individual weights
logP, individual weights
?65
?150
?200
?250
?70
?75
?80
?85
?300
?90
?350
?95
?400
?400
?350
?300
?250
?200
?150
logP, global weights
?100
?50
0
?100
?100
?95
?90
?85
?80
?75
?70
logP, global weights
?65
?60
?55
?50
Figure 5: Scatter plots of the log probability scores for each individual on out-ofsample transactions, plotting log probability scores for individual weights versus log
probability scores for the global weights model. Left: all data, Right: close up.
number of mixture components K for the mixture-based ML weights, the mixturebased Global weights (where all individuals are assigned the same marginal mixture
weights), the mixture-based Empirical Bayes weights, and the non-mixture MAP
histogram method (as a baseline). The mixture-based approaches generally outperform the non-mixture MAP histogram approach (solid line). The ML-based mixture
weights start to overfit after about 6 mixture components (as expected). The Global
mixture weights and individualized mixture weights improve up to about K = 50
components and then show some evidence of overfitting. The mixture-based individual weights method is systematically the best predictor, providing a 15% decrease
in predictive entropy compared to the MAP histogram method, and a roughly 3%
decrease compared to non-individualized global mixture weights.
Figure 5 shows a more detailed comparison of the difference between individual
mixtures and the Global profiles, on a subset of individuals. We can see that the
Global profiles are systematically worse than the individual weights model (i.e., most
points are above the bisecting line). For individuals with the lowest likelihood (lower
left of the left plot) the individual weight model is consistently better: typically
lower weight total likelihood individuals are those with more transactions and items.
In Cadez et al. (2001) we report more detailed results on both this data set and a
second retail data set involving 15 million items and 300,000 individuals. On both
data sets the individual-level models were found to be consistently more accurate
out-of-sample compared to both non-mixture and non-Bayesian approaches. We
also found (empirically) that the time taken for EM to converge is roughly linear
as both a function of number of components and the number of transactions (plots
are omitted due to lack of space), allowing for example fitting of models with 100
mixture components to approximately 2 million baskets in a few hours.
4
Conclusions
In this paper we investigated the use of mixture models and approximate Bayesian
estimation for automatically inferring individual-level profiles from transaction data
records. On a real-world retail data set the proposed framework consistently outperformed alternative approaches in terms of accuracy of predictions on future unseen
customer behavior.
Acknowledgements
The research described in this paper was supported in part by NSF award IRI9703120. The work of Igor Cadez was supported by a Microsoft Graduate Research
Fellowship.
References
Agrawal, R., Imielenski, T., and Swami, A. (1993) Mining association rules between sets
of items in large databases, Proceedings of the ACM SIGMOD Conference on
Management of Data (SIGMOD?98), New York: ACM Press, pp. 207?216.
Cadez, I. V., Smyth, P., Ip, E., Mannila, H. (2001) Predictive profiles for transaction
data using finite mixture models, Technical Report UCI-ICS-01-67, Information and Computer Science, University of California, Irvine (available online at
www.datalab.uci.edu.
Heckerman, D., Chickering, D. M., Meek, C., Rounthwaite, R., and Kadie, C. (2000)
Dependency networks for inference, collaborative filtering, and data visualization.
Journal of Machine Learning Research, 1, pp. 49?75.
Hoffmann, T. (1999) Probabilistic latent sematic indexing, Proceedings of the ACM SIGIR
Conference 1999, New York: ACM Press, 50?57.
Lawrence, R.D., Almasi, G.S., Kotlyar, V., Viveros, M.S., Duri, S.S. (2001) Personalization of supermarket product recommendations, Data Mining and Knowledge
Discovery, 5 (1/2).
Lazarsfeld, P. F. and Henry, N. W. (1968) Latent Structure Analysis, New York:
Houghton Mifflin.
McCallum, A. (1999) Multi-label text classification with a mixture model trained by EM,
in AAAI?99 Workshop on Text Learning.
Strehl, A. and J. Ghosh (2000) Value-based customer grouping from large retail datasets,
Proc. SPIE Conf. on Data Mining and Knowledge Discovery, SPIE Proc. Vol.
4057, Orlando, pp 33?42.
| 2120 |@word solid:1 contains:2 score:10 cadez:7 document:3 tuned:3 past:1 existing:1 scatter:1 subsequent:1 numerical:1 hofmann:1 plot:5 generative:3 selected:3 item:21 mccallum:2 yi1:1 ith:1 record:3 provides:1 location:1 dn:1 ik:10 consists:4 combine:1 fitting:1 manner:1 expected:2 indeed:1 roughly:2 market:2 behavior:7 multi:1 shirt:1 buying:1 automatically:1 little:1 window:1 provided:1 bounded:1 mass:1 lowest:1 interpreted:1 pursue:1 ghosh:2 impractical:1 act:1 tie:1 appear:2 treat:1 approximately:2 might:1 eb:2 graduate:1 commerce:1 practical:1 mannila:1 empirical:5 thought:1 word:1 numbered:1 cannot:1 close:1 www:1 equivalent:2 map:7 customer:9 sigir:1 rule:2 spanned:1 population:6 exploratory:2 traditionally:1 massive:1 smyth:3 engage:1 us:2 particularly:1 houghton:1 database:1 observed:3 enters:1 calculate:1 decrease:2 intuition:1 govern:1 trained:1 swami:2 predictive:14 basis:7 bisecting:1 represented:1 various:1 train:1 describe:1 effective:3 monte:1 lazarsfeld:2 quite:1 whose:1 ability:2 unseen:1 yini:1 think:1 ip:1 online:1 agrawal:2 propose:2 product:4 remainder:1 uci:4 mifflin:1 adapts:1 description:3 sematic:1 generating:2 illustrate:1 derive:1 ij:1 expectationmaximization:1 predicted:1 orlando:1 shopping:1 generalization:1 yij:8 extension:1 clothing:3 ic:3 lawrence:2 predict:1 vary:1 omitted:1 purpose:2 estimation:5 proc:2 outperformed:1 label:1 visited:1 propensity:1 agrees:1 vice:1 reflects:1 hoffman:1 rather:1 focus:2 she:1 consistently:4 bernoulli:1 indicates:1 likelihood:3 baseline:3 posteriori:1 inference:2 typically:2 borrowing:1 kc:3 i1:1 semantics:1 classification:1 flexible:1 denoted:1 proposes:1 smoothing:1 marginal:3 represents:2 broad:2 igor:2 purchase:7 simplex:3 report:2 future:1 few:2 randomly:1 simultaneously:1 individual:82 phase:1 microsoft:1 suit:2 interest:2 mining:4 investigate:1 mixture:51 personalization:2 chain:2 accurate:2 integral:1 necessary:1 plotted:1 nij:2 modeling:7 logp:8 subset:1 predictor:1 reported:1 dependency:1 probabilistic:7 again:1 aaai:1 recorded:1 management:2 padhraic:1 choose:1 woman:1 worse:1 conf:1 kadie:1 coefficient:2 blurred:1 explicitly:1 later:1 tion:1 closed:1 optimistic:1 analyze:1 start:1 bayes:2 collaborative:2 chart:1 ni:1 accuracy:1 correspond:2 bayesian:12 carlo:1 history:1 basket:5 nonetheless:1 pp:3 involved:1 di:3 spie:2 irvine:3 knowledge:2 routine:1 actually:1 appears:1 higher:3 methodology:1 evaluated:1 shrink:1 overfit:2 hand:1 web:3 lack:2 effect:4 assigned:4 memoryless:2 deal:1 conditionally:1 during:2 complete:2 recently:1 multinomial:10 empirically:3 million:2 association:2 discussed:1 occurred:1 he:2 significant:1 versa:1 henry:2 etc:1 own:1 recent:1 store:6 yi:3 scoring:1 seen:1 somewhat:1 freely:1 converge:1 period:6 multiple:1 timestamped:1 smooth:1 technical:1 profiling:1 cross:1 divided:1 visit:2 award:1 prediction:2 involving:1 histogram:9 represent:2 bimodal:1 retail:9 fellowship:1 tend:1 spirit:2 call:3 extracting:1 leverage:1 fit:1 idea:4 prototype:2 six:1 pca:1 forecasting:2 york:3 generally:1 clear:1 involve:1 detailed:3 amount:1 category:2 outperform:2 nsf:1 estimated:3 per:2 discrete:1 vol:1 key:1 year:2 run:1 uncertainty:2 telecommunication:1 striking:1 family:1 draw:1 bit:2 capturing:1 meek:1 strength:2 occur:1 personalized:1 relatively:2 department:10 combination:1 heckerman:2 across:2 em:6 cth:1 making:3 indexing:1 taken:1 computationally:1 visualization:1 describing:1 count:5 discus:1 mechanism:1 available:1 hierarchical:1 generic:1 alternative:1 clustering:1 include:1 dirichlet:3 sigmod:2 k1:1 purchased:8 hoffmann:1 dependence:1 kth:1 individualized:8 separate:1 collected:1 reason:1 providing:2 difficult:1 negative:3 resurgence:1 unknown:2 allowing:1 markov:1 datasets:1 finite:1 displayed:1 heterogeneity:1 smoothed:3 arbitrary:1 inferred:1 pair:1 namely:1 california:3 learned:1 hour:1 address:3 bar:1 below:4 pattern:4 ofsample:1 everyone:1 power:1 event:2 predicting:1 scheme:1 improve:1 numerous:1 categorical:1 rounthwaite:1 kj:1 supermarket:1 text:4 prior:4 literature:1 acknowledgement:1 discovery:2 relative:1 fully:3 expect:1 men:1 limitation:1 filtering:2 versus:1 borrows:1 purchasing:2 principle:1 plotting:2 systematically:2 strehl:2 course:2 token:1 placed:1 supported:2 jth:2 world:3 made:1 adaptive:1 historical:3 social:1 transaction:43 approximate:3 ml:4 global:16 overfitting:1 latent:3 learn:2 ca:2 ignoring:1 investigated:1 did:1 pk:4 profile:27 allowed:1 site:1 representative:1 fashion:2 predictability:1 inferring:2 explicit:1 chickering:1 specific:15 evidence:1 grouping:1 consist:2 exists:1 workshop:1 effectively:1 conditioned:1 smoothly:1 entropy:5 assump:1 recommendation:1 acm:4 goal:1 viewed:2 towards:1 shared:1 telephone:1 except:1 operates:1 determined:2 total:4 experimental:1 indicating:1 evaluate:3 d1:1 handling:1 |
1,231 | 2,121 | Analysis of Sparse Bayesian Learning
Anita C. Fanl
Michael E. Tipping
Microsoft Research
St George House, 1 Guildhall St
Cambridge CB2 3NH, U.K .
Abstract
The recent introduction of the 'relevance vector machine' has effectively demonstrated how sparsity may be obtained in generalised
linear models within a Bayesian framework. Using a particular
form of Gaussian parameter prior, 'learning' is the maximisation,
with respect to hyperparameters, of the marginal likelihood of the
data. This paper studies the properties of that objective function, and demonstrates that conditioned on an individual hyperparameter, the marginal likelihood has a unique maximum which
is computable in closed form. It is further shown that if a derived
'sparsity criterion' is satisfied, this maximum is exactly equivalent
to 'pruning' the corresponding parameter from the model.
1
Introduction
We consider the approximation, from a training sample, of real-valued functions,
a task variously referred to as prediction, regression, interpolation or function approximation. Given a set of data {xn' tn};;=l the 'target' samples tn = f(xn) + En
are conventionally considered to be realisations of a deterministic function f, potentially corrupted by some additive noise process. This function f will be modelled
by a linearly-weighted sum of M fixed basis functions {4>m (X)}~= l:
M
f(x) =
L
wm?>m(x),
(1)
m=l
and the objective is to infer values of the parameters/weights {Wm}~=l such that
is a 'good' approximation of f.
f
While accuracy in function approximation is generally universally valued, there has
been significant recent interest [2, 9, 3, 5]) in the notion of sparsity, a consequence
of learning algorithms which set significant numbers of the parameters Wm to zero.
A methodology which effectively combines both these measures of merit is that of
'sparse Bayesian learning', briefly reviewed in Section 2, and which was the basis for the recent introduction of the relevance vector machine (RVM) and related
models [6, 1, 7]. This model exhibits some very compelling properties, in particular
a dramatic degree of sparseness even in the case of highly overcomplete basis sets
(M ?N). The sparse Bayesian learning algorithm essentially involves the maximisation of a marginalised likelihood function with respect to hyperparameters in
the model prior. In the RVM , this was achieved through re-estimation equations,
the behaviour of which was not fully understood. In this paper we present further
relevant theoretical analysis of the properties of the marginal likelihood which gives
a much fuller picture of the nature of the model and its associated learning procedure. This is detailed in Section 3, and we close with a summary of our findings and
discussion of their implications in Section 4 (and which, to avoid repetition here,
the reader may wish to preview at this point).
2
Sparse Bayesian Learning
We now very briefly review the methodology of sparse Bayesian learning, more
comprehensively described elsewhere [6]. To simplify and generalise the exposition,
we omit to notate any functional dependence on the inputs x and combine quantities
defined over the training set and basis set within N- and M-vectors respectively.
Using this representation, we first write the generative model as:
t = f
+ ?,
(2)
where t = (t1"'" tN )T, f = (11, ... , fN)T and ? = (E1"'" EN)T. The approximator
is then written as:
f = <I>w,
(3)
where <I> = [?Pl'" ?PM] is a general N x M design matrix with column vectors ?Pm
and w = (W1, ... ,WM)T. Recall that in the context of (1), <I>nm = ?m(x n ) and
f
= {f(x1), .. .j(XN)P.
The sparse Bayesian framework assumes an independent zero-mean Gaussian noise
model, with variance u 2 , giving a multivariate Gaussian likelihood of the target
vector t:
p(tlw, ( 2) = (27r)-N/2 U -N exp { _lit
~:"2
}.
(4)
The prior over the parameters is mean-zero Gaussian:
M
p(wlo:) = (27r)-M/21I
a~,e exp
(
- a m2W2)
m
,
(5)
where the key to the model sparsity is the use of M independent hyperparameters
= (a1 " '" aM)T, one per weight (or basis vector), which moderate the strength
of the prior. Given 0:, the posterior parameter distribution is Gaussian and given
via Bayes' rule as p(wlt , 0:) = N(wIIL,~) with
0:
~ =
(A + u - 2<I> T<I? - 1
IL =
u - 2~<I>Tt,
(6)
and A defined as diag(a1, ... ,aM) . Sparse Bayesian learning can then be formulated as a type-II maximum likelihood procedure, in that objective is to maximise
the marginal likelihood, or equivalently, its logarithm ?(0:) with respect to the hyperparameters 0::
?(0:) = logp(tlo: , ( 2) = log
=
with C = u 2I
1
-"2
[Nlog27r
+ <I> A - l<I>T.
i:
p(tlw, ( 2) p(wlo:) dw,
+ log ICI + t C- 1t]
T
,
(7)
(8)
Once most-probable values aMP have been found 1 , in practice they can be plugged
into (6) to give a posterior mean (most probabletpoint estimate for the parameters
J.tMP and from that a mean final approximator: f MP = ()J.tMp?
Empirically, the local maximisation of the marginal likelihood (8) with respect to
a has been seen to work highly effectively [6, 1, 7]. Accurate predictors may be
realised, which are typically highly sparse as a result of the maximising values of
many hyperparameters being infinite. From (6) this leads to a parameter posterior
infinitely peaked at zero for many weights Wm with the consequence that J.tMP
correspondingly comprises very few non-zero elements.
However, the learning procedure in [6] relied upon heuristic re-estimation equations
for the hyperparameters, the behaviour of which was not well characterised. Also,
little was known regarding the properties of (8), the validity of the local maximisation thereof and importantly, and perhaps most interestingly, the conditions under
which a-values would become infinite. We now give, through a judicious re-writing
of (8), a more detailed analysis of the sparse Bayesian learning procedure.
3
3.1
Properties of the Marginal Likelihood ?(0:)
A convenient re-writing
We re-write C from (8) in a convenient form to analyse the dependence on a single
hyperparameter ai:
C = (]"21 +
2..: am?m?~' = (]"21 + 2..: a~1?m?~ + a-;1?i?T,
m
m# i
= C_ i + a-;1?i?T,
(9)
where we have defined C- i = (]"21+ Lm#i a;r/?m?~ as the covariance matrix with
the influence of basis vector ?i removed, equivalent also to ai = 00.
Using established matrix determinant and inverse identities, (9) allows us to write
the terms of interest in ?( a) as:
(10)
(11)
which gives
?(a) =
-~ [Nlog(2n) + log IC-il + tTC=;t
(?TC- 1 t)2
. ? T-' -1 ],
a. + ?i C_ i ?i
1
(?TC- 1 t)2
= ?(a-i) + -2 [logai -log(ai + ?TC=;?i) +
? T-' -1 ]
ai + ?i C_ i ?i
= ?( a-i) + ?( ai),
-logai + log(ai + ?TC=;?i) -
,
(12)
where ?(a-i) is the log marginal likelihood with ai (and thus Wi and ?i) removed
from the model and we have now isolated the terms in ai in the function ?(ai).
IThe most-probable noise variance (]"~p can also be directly and successfully estimated
from the data [6], but for clarity in this paper, we assume without prejudice to our results
that its value is fixed.
3.2
First derivatives of ?(0:)
Previous results. In [6], based on earlier results from [4], the gradient of the
marginal likelihood was computed as:
(13)
with fJi the i-th element of JL and ~ ii the i-th diagonal element of~. This then leads
to re-estimation updates for O::i in terms of fJi and ~ii where, disadvantageously,
these latter terms are themselves functions of O::i.
A new, simplified, expression.
In fact , by instead differentiating (12) directly,
(13) can be seen to be equivalent to:
(14)
where advantageously, O::i now occurs only explicitly since C - i is independent of O::i.
For convenience, we combine terms and re-write (14) as:
o::;lS;- (Qr - Si)
o?(o:) _
OO::i
2( O::i
+ Si)2
(15)
where, for simplification of this and forthcoming expressions, we have defined:
(16)
The term Qi can be interpreted as a 'quality' factor: a measure of how well c/>i
increases ?(0:) by helping to explain the data, while Si is a 'sparsity' factor which
measures how much the inclusion of c/>i serves to decrease ?(0:) through 'inflating'
C (i. e. adding to the normalising factor).
3.3
Stationary points of ?(0:)
Equating (15) to zero indicates that stationary points of the marginal likelihood
occur both at O::i = +00 (note that , being an inverse variance, O::i must be positive)
and for:
. _
O::t subject to Qr
S2t
Q; _Si'
(17)
> Si as a consequence again of O::i > o.
Since the right-hand-side of (17) is independent of O::i, we may find the stationary
points of ?(O::i) analytically without iterative re-estimation. To find the nature of
those stationary points, we consider the second derivatives.
3.4
3.4.1
Second derivatives of ?(0:)
With respect to O::i
Differentiating (15) a second time with respect to O::i gives:
-0::;2S;(O::i + Si)2 - 2(O::i
+ Si) [o::;lS;- (Qr - Si)]
2(O::i + Si)4
and we now consider (18) for both finite- and infinite-O::i stationary points.
(18)
Finite 0::. In this case, for stationary points given by (17), we note that the second
term in the numerator in (18) is zero, giving:
(19)
We see that (19) is always negative, and therefore ?(O::i) has a maximum, which
Si > and O::i given by (17).
must be unique, for
?
Q; -
Infinite 0::. For this case, (18) and indeed, all further derivatives, are uninformatively zero at O::i = 00 , but from (15) we can see that as O::i --+ 00, the sign of the
gradient is given by the sign of - (Q; - Si).
Q; -
If
Si > 0, then the gradient at O::i = 00 is negative so as O::i decreases ?(O::i)
must increase to its unique maximum given by (17). It follows that O::i = 00 is thus
a minimum. Conversely, if
Si < 0, O::i = 00 is the unique maximum of ?(O::i) .
If
Si = 0, then this maximum and that given by (17) coincide.
Q; -
Q; -
We now have a full characterisation of the marginal likelihood as a function of a
single hyperparameter, which is illustrated in Figure 1.
u.
10'
10?
I
Q;
Figure 1: Example plots of ?(ai) against a i (on a log scale) for
> Si (left) ,
showing the single maximum at finite ai, and
< Si (right), showing the
maximum at a i = 00.
Q;
3.4.2
With respect to
O::j,
j
i:- i
To obtain the off-diagonal terms of the second derivative (Hessian) matrix, it is
convenient to manipulate (15) to express it in terms of C. From (11) we see that
and
(20)
Utilising these identities in (15) gives:
(21)
We now write:
(22)
where 6ij is the Kronecker 'delta' function , allowing us to separate out the additional
(diagonal) term that appears only when i = j.
Writing, similarly to (9) earlier, C = C_ j
differentiating with respect to aj gives:
+ ajl?j?j,
substituting into (21) and
while we have
(24)
If all hyperparameters ai are individually set to their maximising values, i. e. a =
aMP such that alI8?(a)/8ai = 0, then even if all 8 2 ?(a)/8a; are negative, there
may still be a non-axial direction in which the likelihood could be increasing. We
now rule out this possibility by showing that the Hessian is negative semi-definite.
First, we note from (21) that if 8?(a)/8ai = 0, 'V;i = 0. Then, if v is a generic
nonzero direction vector:
<
(25)
where we use the Cauchy-Schwarz inequality. If the gradient vanishes, then for all
00, or from (21) , ?rC-1?i = (?rC- 1t)2. It follows
directly from (25) that the Hessian is negative semi-definite, with (25) only zero
where v is orthogonal to all finite a values.
i = 1, ... , M either ai =
4
Summary
Sparse Bayesian learning proposes the iterative maximisation of the marginal likelihood function ?(a) with respect to the hyperparameters a. Our analysis has
shown the following:
1. As a function of an individual hyperparameter ai, ?( a) has a unique maximum
computable in closed-form. (This maximum is, of course, dependent on the
values of all other hyperparameters.)
II. If the criterion
occurs at O:i =
model.
Qr - Si
00,
(defined in Section 3.2) is negative, this maximum
equivalent to the removal of basis function i from the
III. The point where all individual marginal likelihood functions ?(O:i) are maximised is a joint maximum (not necessarily unique) over all O:i.
These results imply the following consequences.
? From I, we see that if we update , in any arbitrary order, the O:i parameters
using (17), we are guaranteed to increase the marginal likelihood at each
step, unless already at a maximum. Furthermore, we would expect these
updates to be more efficient than those given in [6], which individually only
increase, not maximise, ? (O:i) .
? Result III indicates that sequential optimisation of individual O:i cannot
lead to a stationary point from which a joint maximisation over all 0: may
have escaped. (i.e. the stationary point is not a saddle point.)
? The result II confirms the qualitative argument and empirical observation
that many O:i -+ 00 as a result of the optimisation procedure in [6]. The
inevitable implication of finite numerical precision prevented the genuine
sparsity of the model being verified in those earlier simulations.
? We conclude by noting that the maximising hyperparameter solution (17)
remains valid if O:i is already infinite. This means that basis functions not
even in the model can be assessed and their corresponding hyperparameters
updated if desired. So as well as the facility to increase ?(0:) through
the 'pruning' of basis functions if
Si ::::: 0, new basis functions can
be introduced if O:i = 00 but
Si > O. This has highly desirable
computational consequences which we are exploiting to obtain a powerful
'constructive' approximation algorithm [8].
Qr Qr -
References
[1] C. M. Bishop and M. E . Tipping. Variational relevance vector machines. In C. Boutilier
and M. Goldszmidt , editors, Proceedings of th e 16th Conference on Uncertainty in
Artificial Intelligence, pages 46- 53. Morgan Kaufmann , 2000.
[2] S. Chen, D. L. Donoho, and M. A. Saunders. Atomic decomposition by basis pursuit.
Technical Report 479, Department of Statistics, Stanford University, 1995.
[3] Y. Grandvalet. Least absolute shrinkage is equivalent to quadratic penalisation. In
L. Niklasson, M. Boden, and T. Ziemske, editors, Proceedings of th e Eighth International Conference on Artificial N eural Networks (ICANN98) , pages 201- 206. Springer,
1998.
[4] D. J. C. MacKay. Bayesian interpolation. Neural Computation, 4(3):415- 447, 1992.
[5] A. J. Smola, B. Scholkopf, and G. Ratsch . Linear programs for automatic accuracy
control in regression. In Proceedings of th e Ninth Int ernational Conference on Artificial
N eural N etworks, pages 575- 580, 1999.
[6] M. E. Tipping. The Relevance Vector Machine. In S. A. Solla, T . K. Leen , and K.-R.
Muller, editors, Advances in N eural Information Processing Systems 12, pages 652- 658.
MIT Press, 2000.
[7] M. E. Tipping. Sparse kernel principal component analysis. In Advances in Neural
Information Processing Systems 13. MIT Press, 200l.
[8] M. E . Tipping and A. C. Faul. Bayesian pursuit. Submitted to NIPS*Ol.
[9] V. N. Vapnik. Statistical Learning Theory. Wiley, 1998.
| 2121 |@word determinant:1 briefly:2 confirms:1 simulation:1 covariance:1 decomposition:1 dramatic:1 tlo:1 interestingly:1 amp:2 si:19 written:1 must:3 fn:1 numerical:1 additive:1 plot:1 update:3 stationary:8 generative:1 intelligence:1 maximised:1 normalising:1 rc:2 become:1 scholkopf:1 s2t:1 qualitative:1 combine:3 indeed:1 themselves:1 ol:1 little:1 increasing:1 interpreted:1 finding:1 inflating:1 exactly:1 demonstrates:1 control:1 omit:1 positive:1 generalised:1 t1:1 understood:1 maximise:2 local:2 consequence:5 interpolation:2 equating:1 conversely:1 unique:6 atomic:1 practice:1 maximisation:6 definite:2 cb2:1 procedure:5 empirical:1 convenient:3 convenience:1 close:1 cannot:1 context:1 influence:1 writing:3 equivalent:5 deterministic:1 demonstrated:1 l:2 rule:2 importantly:1 dw:1 notion:1 updated:1 target:2 element:3 solla:1 decrease:2 removed:2 vanishes:1 ithe:1 upon:1 basis:11 joint:2 artificial:3 saunders:1 heuristic:1 stanford:1 valued:2 tlw:2 statistic:1 analyse:1 final:1 nlog:1 relevant:1 qr:6 exploiting:1 oo:1 axial:1 ij:1 involves:1 faul:1 direction:2 behaviour:2 probable:2 pl:1 c_:4 helping:1 considered:1 ic:1 guildhall:1 exp:2 tmp:3 lm:1 substituting:1 estimation:4 rvm:2 schwarz:1 individually:2 repetition:1 successfully:1 weighted:1 mit:2 gaussian:5 always:1 avoid:1 shrinkage:1 derived:1 likelihood:17 indicates:2 am:3 dependent:1 anita:1 typically:1 proposes:1 mackay:1 marginal:13 genuine:1 once:1 fuller:1 lit:1 inevitable:1 peaked:1 wlt:1 report:1 simplify:1 realisation:1 few:1 individual:4 variously:1 microsoft:1 preview:1 interest:2 highly:4 possibility:1 ziemske:1 implication:2 accurate:1 orthogonal:1 unless:1 plugged:1 logarithm:1 re:8 desired:1 overcomplete:1 isolated:1 theoretical:1 column:1 earlier:3 compelling:1 logp:1 predictor:1 corrupted:1 notate:1 st:2 international:1 off:1 michael:1 w1:1 again:1 satisfied:1 nm:1 derivative:5 int:1 mp:1 explicitly:1 closed:2 realised:1 wm:5 bayes:1 relied:1 il:2 accuracy:2 variance:3 kaufmann:1 modelled:1 bayesian:12 submitted:1 explain:1 against:1 thereof:1 associated:1 wlo:2 recall:1 appears:1 tipping:5 methodology:2 leen:1 furthermore:1 smola:1 hand:1 aj:1 quality:1 perhaps:1 validity:1 facility:1 analytically:1 nonzero:1 illustrated:1 numerator:1 criterion:2 tt:1 tn:3 variational:1 niklasson:1 functional:1 empirically:1 nh:1 jl:1 significant:2 cambridge:1 ai:16 automatic:1 pm:2 similarly:1 inclusion:1 multivariate:1 posterior:3 recent:3 moderate:1 ajl:1 inequality:1 muller:1 seen:2 minimum:1 george:1 additional:1 morgan:1 ii:5 semi:2 full:1 desirable:1 infer:1 technical:1 wiil:1 escaped:1 penalisation:1 e1:1 manipulate:1 prevented:1 a1:2 qi:1 prediction:1 regression:2 optimisation:2 essentially:1 kernel:1 achieved:1 ratsch:1 subject:1 noting:1 iii:2 forthcoming:1 regarding:1 computable:2 expression:2 advantageously:1 hessian:3 boutilier:1 generally:1 detailed:2 sign:2 estimated:1 delta:1 per:1 write:5 hyperparameter:5 express:1 key:1 characterisation:1 clarity:1 verified:1 sum:1 utilising:1 inverse:2 powerful:1 uncertainty:1 reader:1 guaranteed:1 simplification:1 quadratic:1 strength:1 occur:1 kronecker:1 argument:1 department:1 wi:1 equation:2 remains:1 etworks:1 merit:1 serf:1 fji:2 pursuit:2 generic:1 assumes:1 giving:2 objective:3 already:2 quantity:1 occurs:2 dependence:2 diagonal:3 exhibit:1 gradient:4 separate:1 cauchy:1 maximising:3 ttc:1 equivalently:1 potentially:1 negative:6 design:1 allowing:1 observation:1 finite:5 ninth:1 arbitrary:1 introduced:1 established:1 nip:1 eighth:1 sparsity:6 program:1 marginalised:1 imply:1 picture:1 conventionally:1 prior:4 review:1 removal:1 fully:1 expect:1 approximator:2 degree:1 editor:3 grandvalet:1 elsewhere:1 summary:2 course:1 side:1 generalise:1 comprehensively:1 correspondingly:1 differentiating:3 absolute:1 sparse:11 xn:3 valid:1 universally:1 ici:1 simplified:1 coincide:1 pruning:2 conclude:1 iterative:2 reviewed:1 nature:2 boden:1 necessarily:1 diag:1 linearly:1 noise:3 hyperparameters:10 x1:1 eural:3 referred:1 en:2 wiley:1 precision:1 comprises:1 wish:1 house:1 bishop:1 showing:3 vapnik:1 adding:1 effectively:3 sequential:1 conditioned:1 sparseness:1 chen:1 tc:4 saddle:1 infinitely:1 springer:1 identity:2 formulated:1 donoho:1 exposition:1 judicious:1 infinite:5 characterised:1 prejudice:1 principal:1 latter:1 goldszmidt:1 assessed:1 relevance:4 constructive:1 |
1,232 | 2,122 | Reducing multiclass to binary by
coupling probability estimates
Bianca Zadrozny
Department of Computer Science and Engineering
University of California, San Diego
La Jolla, CA 92093-0114
[email protected]
Abstract
This paper presents a method for obtaining class membership probability estimates for multiclass classification problems by coupling the probability estimates
produced by binary classifiers. This is an extension for arbitrary code matrices
of a method due to Hastie and Tibshirani for pairwise coupling of probability
estimates. Experimental results with Boosted Naive Bayes show that our method
produces calibrated class membership probability estimates, while having similar
classification accuracy as loss-based decoding, a method for obtaining the most
likely class that does not generate probability estimates.
1 Introduction
The two most well-known approaches for reducing a multiclass classification problem to
a set of binary classification problems are known as one-against-all and all-pairs. In the
one-against-all approach, we train a classifier for each of the classes using as positive examples the training examples that belong to that class, and as negatives all the other training
examples. In the all-pairs approach, we train a classifier for each possible pair of classes
ignoring the examples that do not belong to the classes in question.
Although these two approaches are the most obvious, Allwein et al. [Allwein et al., 2000]
have shown that there are many other ways in which a multiclass problem can be decomposed into a number of binary classification problems. We can represent each such decom1 0 1 k l , where k is the number of classes and l is
position by a code matrix M
the number of binary classification problems. If M c b
1 then the examples belonging to class c are considered to be positive examples for the binary classification problem
b. Similarly, if M c b
1 the examples belonging to c are considered to be negative
0 the examples belonging to c are not used in training
examples for b. Finally, if M c b
a classifier for b.
For example, in the 3-class case, the all-pairs code matrix is
c1
c2
c3
b1
1
1
0
b2
1
0
1
b3
0
1
1
This approach for representing the decomposition of a multiclass problem into binary prob-
lems is a generalization of the Error-Correcting Output Codes (ECOC) scheme proposed
by Dietterich and Bakiri [Dietterich and Bakiri, 1995]. The ECOC scheme does not allow
zeros in the code matrix, meaning that all examples are used in each binary classification
problem.
Orthogonal to the problem of choosing a code matrix for reducing multiclass to binary is
the problem of classifying an example given the labels assigned by each binary classifier.
Given an example x, Allwein et al. [Allwein et al., 2000] first create a vector v of length l
containing the -1,+1 labels assigned to x by each binary classifier. Then, they compute
the Hamming distance between v and each row of M, and find the row c that is closest to v
according to this metric. The label c is then assigned to x. This method is called Hamming
decoding.
For the case in which the binary classifiers output a score whose magnitude is a measure of confidence in the prediction, they use a loss-based decoding approach that takes
into account the scores to calculate the distance between v and each row of M, instead
of using the Hamming distance. This method is called loss-based decoding. Allwein et
al. [Allwein et al., 2000] present theoretical and experimental results indicating that this
method is better than Hamming decoding.
However, both of these methods simply assign a class label to each example. They do not
output class membership probability estimates P? C c X x for an example x. These
probability estimates are important when the classification outputs are not used in isolation
and must be combined with other sources of information, such as misclassification costs
[Zadrozny and Elkan, 2001a] or the outputs of another classifier.
Given a code matrix M and a binary classification learning algorithm that outputs probability estimates, we would like to couple the estimates given by each binary classifier in order
to obtain class probability membership estimates for the multiclass problem.
Hastie and Tibshirani [Hastie and Tibshirani, 1998] describe a solution for obtaining probability estimates P? C c X x in the all-pairs case by coupling the pairwise probability
estimates, which we describe in Section 2. In Section 3, we extend the method to arbitrary
code matrices. In Section 4 we discuss the loss-based decoding approach in more detail
and compare it mathematically to the method by Hastie and Tibshirani. In Section 5 we
present experimental results.
2 Coupling pairwise probability estimates
We are given pairwise probability estimates ri j x for every class i j, obtained by training
a classifier using the examples belonging to class i as positives and the examples belonging
to class j as negatives. We would like to couple these estimates to obtain a set of class
membership probabilities pi x
P C ci X x for each example x. The ri j are related
to the pi according to
ri j x P C i C i C
j X
x
pi x
pi x p j x
Since we additionally require that ?i pi x
1, there are k 1 free parameters and k k
1
2 constraints. This implies that there may not exist pi satisfying these constraints.
Let ni j be the number of training examples used to train the binary classifier that predicts
p?i x
p?i x
p? j x , Hastie and
ri j . In order to find the best approximation r?i j x
Tibshirani fit the Bradley-Terrey model for paired comparisons [Bradley and Terry, 1952]
by minimizing the average weighted Kullback-Leibler distance l x between r i j x and
r?i j x for each x, given by
l x
x
1
ij
ni j ri j x log
?
?ri j x
i j
r
ri j x log
1
1
ri j x
?ri j x
The algorithm is as follows:
1. Start with some guess for the ?pi x and corresponding ?ri j x .
2. Repeat until convergence:
(a) For each i 1 2 k
?pi x
? j i n i j ri j x
? j i ni j ?ri j x
?pi x
(b) Renormalize the ?pi x .
(c) Recompute the ?ri j x .
Hastie and Tibshirani [Hastie and Tibshirani, 1998] prove that the Kullback-Leibler distance between ri j x and r?i j x decreases at each step. Since this distance is bounded below
by zero, the algorithm converges. At convergence, the r?i j are consistent with the p? i . The
class predicted for each example x is c? x
argmax p? i x .
Hastie and Tibshirani also prove that the p? i x are in the same order as the non-iterative
estimates p?i x
? j i ri j x for each x. Thus, the p?i x are sufficient for predicting the
most likely class for each example. However, as shown by Hastie and Tibshirani, they
are not accurate probability estimates because they tend to underestimate the differences
between the p?i x values.
3 Extending the Hastie-Tibshirani method to arbitrary code matrices
For an arbitrary code matrix M, instead of having pairwise probability estimates, we have
an estimate rb x for each column b of M, such that
rb x P
C
c I
c
c I J
C c X
? c I pc x
? c I J pc x
x
1, respectively.
We would like to obtain a set of class membership probabilities pi
x for each example
x compatible with the rb
x and subject to ?i pi
x 1. In this case, the number of free
parameters is k 1 and the number of constraints is l 1, where l is the number of columns
of the code matrix.
Since for most code matrices l is greater than k 1, in general there is no exact solution to
where I and J are the set of classes for which M b
1 and M b
this problem. For this reason, we propose an algorithm analogous to the Hastie-Tibshirani
method presented in the previous section to find the best approximate probability estimates
p?i (x) such that
?rb x
?c I ?pc x
?c I J ?pc x
and the Kullback-Leibler distance between r?b x and rb x is minimized.
Let nb be the number of training examples used to train the binary classifier that corresponds to column b of the code matrix. The algorithm is as follows:
1. Start with some guess for the ?pi x and corresponding ?rb x .
2. Repeat until convergence:
(a) For each i 1 2 k
?pi x
?pi x
?b s t
?b s t
1 n b rb x
rb x
M i b 1 nb ?
M ib
?b s t
?b s t
1 nb 1 rb x
rb x
M i b 1 nb 1 ?
M ib
(b) Renormalize the ?pi x .
(c) Recompute the ?rb x .
If the code matrix is the all-pairs matrix, this algorithm reduces to the original method by
Hastie and Tibshirani.
1 and B i be the set of matrix
Let B i be the set of matrix columns for which M i
columns for which M c
1. By analogy with the non-iterative estimates suggested
by Hastie and Tibshirani, we can define non-iterative estimates p?i x
? b B
i rb x
?b B i 1 rb x . For the all-pairs code matrix, these estimates are the same as the ones
suggested by Hastie and Tibshirani. However, for arbitrary matrices, we cannot prove that
the non-iterative estimates predict the same class as the iterative estimates.
4 Loss-based decoding
In this section, we discuss how to apply the loss-based decoding method to classifiers that
output class membership probability estimates. We also study the conditions under which
this method predicts the same class as the Hastie-Tibshirani method, in the all-pairs case.
The loss-based decoding method [Allwein et al., 2000] requires that each binary classifier
output a margin score satisfying two requirements. First, the score should be positive if
the example is classified as positive, and negative if the example is classified as negative.
Second, the magnitude of the score should be a measure of confidence in the prediction.
The method works as follows. Let f x b be the margin score predicted by the classifier
corresponding to column b of the code matrix for example x. For each row c of the code
matrix M and for each example x, we compute the distance between f and M c as
dL x c
l
L M c b f x b
?
(1)
b 1
where L is a loss function that is dependent on the nature of the binary classifier and M c b
= 0, 1 or 1. We then label each example x with the label c for which dL is minimized.
If the binary classification learning algorithm outputs scores that are probability estimates,
they do not satisfy the first requirement because the probability estimates are all between 0
and 1. However, we can transform the probability estimates rb x output by each classifier
b into margin scores by subtracting 1
2 from the scores, so that we consider as positives
the examples x for which rb x is above 1/2, and as negatives the examples x for which
rb x is below 1/2.
We now prove a theorem that relates the loss-based decoding method to the HastieTibshirani method, for a particular class of loss functions.
Theorem 1 The loss-based decoding method for all-pairs code matrices predicts the same
class label as the iterative estimates p? i x given by Hastie and Tibshirani, if the loss function
is of the form L y
ay, for any a
0.
Proof: We first show that, if the loss function is of the form L y
ay, the loss-based
decoding method predicts the same class label as the non-iterative estimates p? i x , for the
all-pairs code matrix.
Dataset
satimage
pendigits
soybean
#Training Examples
4435
7494
307
#Test Examples
2000
3498
376
#Attributes
36
16
35
#Classes
7
10
9
Table 1: Characteristics of the datasets used in the experiments.
The non-iterative estimates p?i x are given by
?pc x
?
rb x
b B
c
? 1
b B
rb x
?
rb x
b B
c
c
?
b B
rb x
c
B c
B c and B c are the sets of matrix columns for which M
c 1 and M
c
where
1, respectively.
ay and f
x b rb
x 1
2, and eliminating the terms for which
Considering that L
y
M
c b 0, we can rewrite Equation 1 as
d x c
?
b B
c
a rb x
1
2
? a rb x
b B
c
1
2
a
?
b B
rb x
c
?
b B
rb x
c
For the all-pairs code matrix the following relationship holds: 1
2 B c
k 1
2, where k is the number of classes. So, the distance d x c is
d x c
a
?
b B
rb x
c
?
b B
rb x
c
B c
k
1
B
2
B
c
c
B
B c
c
1 2
It is now easy to see that the class c x which minimizes d x c for example x, also maximizes p?c x . Furthermore, if d x i d x j then p x i
p x j , which means that the
ranking of the classes for each example is the same.
Since the non-iterative estimates p? c x are in the same order as the iterative estimates p? c x ,
we can conclude that the Hastie-Tibshirani method is equivalent to the loss-based decoding
method if L y
ay, in terms of class prediction, for the all-pairs code matrix.
Allwein et al. do not consider loss functions of the form L y
ay, and uses non-linear
e y . In this case, the class predicted by loss-based decoding
loss functions such as L y
may differ from the one predicted by the method by Hastie and Tibshirani.
This theorem applies only to the all-pairs code matrix. For other matrices such that
B c
B c is a linear function of B c (such as the one-against-all matrix), we can prove
ay) predicts the same class as the non-iterative esthat loss-based decoding (with L y
timates. However, in this case, the non-iterative estimates do not necessarily predict the
same class as the iterative ones.
5 Experiments
We performed experiments using the following multiclass datasets from the UCI Machine
Learning Repository [Blake and Merz, 1998]: satimage, pendigits and soybean. Table
1 summarizes the characteristics of each dataset.
The binary learning algorithm used in the experiments is boosted naive Bayes
[Elkan, 1997], since this is a method that cannot be easily extended to handle multiclass
problems directly. For all the experiments, we ran 10 rounds of boosting.
Method
Loss-based (L y y)
Loss-based (L y e y )
Hastie-Tibshirani (non-iterative)
Hastie-Tibshirani (iterative)
Loss-based (L y y)
Loss-based (L y e y )
Extended Hastie-Tibshirani (non-iterative)
Extended Hastie-Tibshirani (iterative)
Loss-based (L y y)
Loss-based (L y e y )
Extended Hastie-Tibshirani (non-iterative)
Extended Hastie-Tibshirani (iterative)
Multiclass Naive Bayes
Code Matrix
All-pairs
All-pairs
All-pairs
All-pairs
One-against-all
One-against-all
One-against-all
One-against-all
Sparse
Sparse
Sparse
Sparse
-
Error Rate
0.1385
0.1385
0.1385
0.1385
0.1445
0.1425
0.1445
0.1670
0.1435
0.1425
0.1480
0.1330
0.2040
MSE
0.0999
0.0395
0.1212
0.0396
0.1085
0.0340
0.0651
Table 2: Test set results on the satimage dataset.
We use three different code matrices for each dataset: all-pairs, one-against-all and a sparse
random matrix. The sparse random matrices have 15 log2 k columns, and each element
is 0 with probability 1/2 and -1 or +1 with probability 1/4 each. This is the same type of
sparse random matrix used by Allwein et al.[Allwein et al., 2000]. In order to have good
error correcting properties, the Hamming distance ? between each pair of rows in the matrix
must be large. We select the matrix by generating 10,000 random matrices and selecting
the one for which ? is maximized, checking that each column has at least one 1 and one
1, and that the matrix does not have two identical columns.
We evaluate the performance of each method using two metrics. The first metric is the
error rate obtained when we assign each example to the most likely class predicted by
the method. This metric is sufficient if we are only interested in classifying the examples
correctly and do not need accurate probability estimates of class membership.
The second metric is squared error, defined for one example x as SE x
?j tj x
2
p j x , where p j x is the probability estimated by the method for example x and class
j, and t j x is the true probability of class j for x. Since for most real-world datasets true
labels are known, but not probabilities, t j x is defined to be 1 if the label of x is j and
0 otherwise. We calculate the squared error for each x to obtain the mean squared error (MSE). The mean squared error is an adequate metrics for assessing the accuracy of
probability estimates [Zadrozny and Elkan, 2001b]. This metric cannot be applied to the
loss-based decoding method, since it does not produce probability estimates.
Table 2 shows the results of the experiments on the satimage dataset for each type of code
matrix. As a baseline for comparison, we also show the results of applying multiclass Naive
Bayes to this dataset. We can see that the iterative Hastie-Tibshirani procedure (and its
extension to arbitrary code matrices) succeeds in lowering the MSE significantly compared
to the non-iterative estimates, which indicates that it produces probability estimates that
are more accurate. In terms of error rate, the differences between methods are small. For
one-against-all matrices, the iterative method performs consistently worse, while for sparse
random matrices, it performs consistently better. Figure 1 shows how the MSE is lowered
at each iteration of the Hastie-Tibshirani algorithm, for the three types of code matrices.
Table 3 shows the results of the same experiments on the datasets pendigits and soybean.
Again, the MSE is significantly lowered by the iterative procedure, in all cases. For the
soybean dataset, using the sparse random matrix, the iterative method again has a lower
error rate than the other methods, which is even lower than the error rate using the all-pairs
matrix. This is an interesting result, since in this case the all-pairs matrix has 171 columns
(corresponding to 171 classifiers), while the sparse matrix has only 64 columns.
Satimage
0.12
0.11
all?pairs
one?against?all
sparse
0.1
0.09
MSE
0.08
0.07
0.06
0.05
0.04
0.03
0
5
10
15
20
25
30
35
Iteration
Figure 1: Convergence of the MSE for the satimage dataset.
Method
Loss-based (L y y)
Loss-based (L y e y )
Hastie-Tibshirani (non-iterative)
Hastie-Tibshirani (iterative)
Loss-based (L y y)
Loss-based (L y e y )
Ext. Hastie-Tibshirani (non-it.)
Ext. Hastie-Tibshirani (it.)
Loss-based (L y y)
Loss-based (L y e y )
Ext. Hastie-Tibshirani (non-it.)
Ext. Hastie-Tibshirani (it.)
Multiclass Naive Bayes
Code Matrix
All-pairs
All-pairs
All-pairs
All-pairs
One-against-all
One-against-all
One-against-all
One-against-all
Sparse
Sparse
Sparse
Sparse
-
pendigits
Error Rate
MSE
0.0723
0.0715
0.0723
0.0747
0.0718
0.0129
0.0963
0.0963
0.0963
0.0862
0.1023
0.0160
0.1284
0.1266
0.1484
0.0789
0.1261
0.0216
0.2779
0.0509
soybean
Error Rate
MSE
0.0665
0.0665
0.0665
0.0454
0.0665
0.0066
0.0824
0.0931
0.0824
0.0493
0.0931
0.0073
0.0718
0.0718
0.0798
0.0463
0.0636
0.0062
0.0745
0.0996
Table 3: Test set results on the pendigits and soybean datasets.
6 Conclusions
We have presented a method for producing class membership probability estimates for
multiclass problems, given probability estimates for a series of binary problems determined
by an arbitrary code matrix.
Since research in designing optimal code matrices is still on-going
[Utschick and Weichselberger, 2001] [Crammer and Singer, 2000], it is important to
be able to obtain class membership probability estimates from arbitrary code matrices.
In current research, the effectiveness of a code matrix is determined primarily by the
classification accuracy. However, since many applications require accurate class membership probability estimates for each of the classes, it is important to also compare the
different types of code matrices according to their ability of producing such estimates. Our
extension of Hastie and Tibshirani?s method is useful for this purpose.
Our method relies on the probability estimates given by the binary classifiers to produce the
multiclass probability estimates. However, the probability estimates produced by Boosted
Naive Bayes are not calibrated probability estimates. An interesting direction for future
work is in determining whether the calibration of the probability estimates given by the
binary classifiers improves the calibration of the multiclass probabilities.
References
[Allwein et al., 2000] Allwein, E. L., Schapire, R. E., and Singer, Y. (2000). Reducing multiclass
to binary: A unifying approach for margin classifiers. Journal of Machine Learning Research,
1:113?141.
[Blake and Merz, 1998] Blake, C. L. and Merz, C. J. (1998). UCI repository of machine learning
databases. Department of Information and Computer Sciences, University of California, Irvine.
http://www.ics.uci.edu/ mlearn/MLRepository.html.
[Bradley and Terry, 1952] Bradley, R. and Terry, M. (1952). Rank analysis of incomplete block
designs, I: The method of paired comparisons. Biometrics, pages 324?345.
[Crammer and Singer, 2000] Crammer, K. and Singer, Y. (2000). On the learnability and design of
output codes for multiclass problems. In Proceedings of the Thirteenth Annual Conference on
Computational Learning Theory, pages 35?46.
[Dietterich and Bakiri, 1995] Dietterich, T. G. and Bakiri, G. (1995). Solving multiclass learning
problems via error-correcting output codes. Journal of Artificial Intelligence Research, 2:263?
286.
[Elkan, 1997] Elkan, C. (1997). Boosting and naive bayesian learning. Technical Report CS97-557,
University of California, San Diego.
[Hastie and Tibshirani, 1998] Hastie, T. and Tibshirani, R. (1998). Classification by pairwise coupling. In Advances in Neural Information Processing Systems, volume 10. MIT Press.
[Utschick and Weichselberger, 2001] Utschick, W. and Weichselberger, W. (2001). Stochastic organization of output codes in multiclass learning problems. Neural Computation, 13(5):1065?1102.
[Zadrozny and Elkan, 2001a] Zadrozny, B. and Elkan, C. (2001a). Learning and making decisions
when costs and probabilities are both unknown. In Proceedings of the Seventh International
Conference on Knowledge Discovery and Data Mining, pages 204?213. ACM Press.
[Zadrozny and Elkan, 2001b] Zadrozny, B. and Elkan, C. (2001b). Obtaining calibrated probability
estimates from decision trees and naive bayesian classifiers. In Proceedings of the Eighteenth
International Conference on Machine Learning, pages 609?616. Morgan Kaufmann Publishers,
Inc.
| 2122 |@word dietterich:4 repository:2 c:1 eliminating:1 predicted:5 implies:1 differ:1 direction:1 true:2 assigned:3 bakiri:4 leibler:3 attribute:1 question:1 stochastic:1 decomposition:1 round:1 distance:10 require:2 mlrepository:1 assign:2 series:1 score:9 selecting:1 generalization:1 ay:6 evaluate:1 reason:1 mathematically:1 performs:2 extension:3 bradley:4 current:1 hold:1 code:36 considered:2 blake:3 ic:1 meaning:1 length:1 must:2 relationship:1 predict:2 minimizing:1 negative:6 purpose:1 volume:1 belong:2 extend:1 design:2 intelligence:1 unknown:1 guess:2 label:10 datasets:5 create:1 weighted:1 similarly:1 zadrozny:8 recompute:2 boosting:2 mit:1 extended:5 ucsd:1 lowered:2 calibration:2 arbitrary:8 boosted:3 c2:1 allwein:12 closest:1 prove:5 pair:26 c3:1 jolla:1 consistently:2 rank:1 indicates:1 california:3 pairwise:6 weichselberger:3 binary:24 baseline:1 ecoc:2 dependent:1 membership:11 morgan:1 utschick:3 decomposed:1 greater:1 able:1 suggested:2 below:2 considering:1 going:1 interested:1 bounded:1 relates:1 maximizes:1 classification:13 html:1 reduces:1 terry:3 misclassification:1 technical:1 minimizes:1 predicting:1 representing:1 scheme:2 having:2 paired:2 identical:1 every:1 prediction:3 metric:7 future:1 minimized:2 classifier:22 report:1 iteration:2 represent:1 primarily:1 naive:8 discovery:1 producing:2 c1:1 checking:1 positive:6 determining:1 engineering:1 thirteenth:1 loss:32 argmax:1 source:1 ext:4 publisher:1 interesting:2 analogy:1 organization:1 subject:1 tend:1 mining:1 pendigits:5 sufficient:2 consistent:1 effectiveness:1 classifying:2 pc:5 pi:16 tj:1 easy:1 row:5 compatible:1 accurate:4 isolation:1 fit:1 hastie:35 repeat:2 block:1 free:2 allow:1 orthogonal:1 procedure:2 biometrics:1 tree:1 incomplete:1 multiclass:19 whether:1 sparse:15 renormalize:2 significantly:2 theoretical:1 world:1 confidence:2 column:12 san:2 cannot:3 nb:4 adequate:1 applying:1 cost:2 useful:1 se:1 www:1 equivalent:1 approximate:1 kullback:3 eighteenth:1 seventh:1 learnability:1 b1:1 conclude:1 generate:1 schapire:1 http:1 correcting:3 exist:1 calibrated:3 combined:1 iterative:26 international:2 estimated:1 tibshirani:35 rb:27 correctly:1 table:6 additionally:1 handle:1 decoding:16 nature:1 ca:1 analogous:1 ignoring:1 obtaining:4 diego:2 squared:4 exact:1 again:2 mse:9 us:1 designing:1 containing:1 necessarily:1 elkan:9 soybean:6 element:1 worse:1 satisfying:2 lowering:1 timates:1 predicts:5 database:1 account:1 prob:1 b2:1 calculate:2 inc:1 bianca:1 satisfy:1 ranking:1 decrease:1 performed:1 decision:2 summarizes:1 ran:1 position:1 start:2 bayes:6 ib:2 theorem:3 annual:1 rewrite:1 solving:1 ni:3 accuracy:3 constraint:3 kaufmann:1 characteristic:2 maximized:1 ri:15 dl:2 easily:1 bayesian:2 produced:2 ci:1 magnitude:2 train:4 describe:2 department:2 classified:2 according:3 artificial:1 mlearn:1 margin:4 choosing:1 belonging:5 simply:1 likely:3 whose:1 against:14 underestimate:1 otherwise:1 making:1 obvious:1 ability:1 lem:1 proof:1 hamming:5 transform:1 couple:2 irvine:1 dataset:8 applies:1 corresponds:1 relies:1 equation:1 acm:1 knowledge:1 improves:1 discus:2 propose:1 subtracting:1 singer:4 satimage:6 uci:3 determined:2 reducing:4 apply:1 called:2 furthermore:1 experimental:3 la:1 merz:3 succeeds:1 until:2 convergence:4 requirement:2 extending:1 assessing:1 produce:4 generating:1 original:1 converges:1 indicating:1 select:1 crammer:3 coupling:6 log2:1 unifying:1 ij:1 b3:1 |
1,233 | 2,123 | Asymptotic Universality for Learning
Curves of Support Vector Machines
M.Opperl
R. Urbanczik 2
1 Neural Computing Research Group
School of Engineering and Applied Science
Aston University, Birmingham B4 7ET, UK.
[email protected]
2Institut Fur Theoretische Physik,
Universitiit Wurzburg Am Rubland, D-97074 Wurzburg, Germany
[email protected].
Abstract
Using methods of Statistical Physics, we investigate the rOle
of model complexity in learning with support vector machines
(SVMs). We show the advantages of using SVMs with kernels
of infinite complexity on noisy target rules, which, in contrast to
common theoretical beliefs, are found to achieve optimal generalization error although the training error does not converge to the
generalization error. Moreover, we find a universal asymptotics of
the learning curves which only depend on the target rule but not
on the SVM kernel.
1
Introduction
Powerful systems for data inference, like neural networks implement complex inputoutput relations by learning from example data. The price one has to pay for the
flexibility of these models is the need to choose the proper model complexity for a
given task, i.e. the system architecture which gives good generalization ability for
novel data. This has become an important problem also for support vector machines
[1]. The main advantage of SVMs is that the learning task is a convex optimization
problem which can be reliably solved even when the example data require the fitting
of a very complicated function. A common argument in computational learning
theory suggests that it is dangerous to utilize the full flexibility of the SVM to learn
the training data perfectly when these contain an amount of noise. By fitting more
and more noisy data, the machine may implement a rapidly oscillating function
rather than the smooth mapping which characterizes most practical learning tasks.
Its prediction ability could be no better than random guessing in that case. Rence,
modifications of SVM training [2] that allow for training errors were suggested to be
necessary for realistic noisy scenarios. This has the drawback of introducing extra
model parameters and spoils much of the original elegance of SVMs.
Surprisingly, the results of this paper show that the picture is rather different in
the important case of high dimensional data spaces. Using methods of Statistical
Physics, we show that asymptotically, the SVM achieves optimal generalization
ability for noisy data already for zero training error. Moreover, the asymptotic rate
of decay of the generalization error is universal, i.e. independent of the kernel that
defines the SVM. These results have been published previously only in a physics
journal [3].
As is well known, SVMs classify inputs y using a nonlinear mapping into a feature
vector w(y) which is an element of a Hilbert space. Based on a training set of m
inputs xl' and their desired classifications 71' , SVMs construct the maximum margin
hyperplane P in the feature space. P can be expressed as a linear combination of
the feature vectors w(xl'), and to classify an input y, that is to decide on which side
of P the image W(y) lies, one basically has to evaluate inner products W(xl') . W(y).
For carefully chosen mappings wand Hilbert spaces, inner products w(x) . w(y)
can be evaluated efficiently using a kernel function k(x, y) = w(x) . w(y), without
having to individually calculate the feature vectors w(x) and w(y). In this manner
it becomes computationally feasible to use very high and even infinite dimensional
feature vectors.
This raises the intriguing question whether the use of a very high dimensional
feature space may typically be helpful. So far, recent results [4, 5] obtained by
using methods of Statistical Mechanics (which are naturally well suited for analysing
stochastic models in high dimensional spaces), have been largely negative in this
respect. They suggest (as one might perhaps expect) that it is rather important to
match the complexity of the kernel to the target rule. The analysis in [4] considers
the case of N-dimensional inputs with binary components and assumes that the
target rule giving the correct classification 7 of an input x is obtained as the sign
of a function t(x) which is polynomial in the input components and of degree L.
The SVM uses a kernel which is a polynomial of the inner product x . y in input
space of degree K ;::: L, and the feature space dimension is thus O(N K ). In this
scenario it is shown, under mild regularity condition on the kernel and for large N,
that the SVM generalizes well when the number of training examples m is on the
order of N L . So the scale of the learning curve is determined by the complexity
of the target rule and not by the kernel. However, considering the rate with which
the generalization error approaches zero one finds the optimal N L 1m decay only
when K is equal to L and the convergence is substantially slower when K > L. So
it is important to match the complexity of the kernel to the target rule and using
a large value of K is only justified if L is assumed large and if one can use on the
order of N L examples for training.
In this paper we show that the situation is very different when one considers the
arguably more realistic scenario of a target rule corrupted by noise. Now one can no
longer use K = L since no separating hyperplane P will exist when m is sufficiently
large compared to N L. However when K > L, this plane will exist and we will show
that it achieves optimal generalization performance in the limit that N L 1m is small.
Remarkably, the asymptotic rate of decay of the generalization error is independent
of the kernel in this case and a general characterization of the asymptote in terms of
properties of the target rule is possible. In a second step we show that under mild
regularity conditions these findings also hold when k(x, y) is an arbitrary function
of x . y or when the kernel is a function of the Euclidean distance Ix - YI. The latter
type of kernels is widely used in practical applications of SVMs.
2
Learning with Noise: Polynomial Kernels
We begin by assuming a polynomial kernel k(x, y) = J(x? y) where J(z) =
Lf=oCk zk is of degree K. Denoting by P a multi-index P = (PI , ... ,PN) with Pi E
No, we set xp =
JTPTfnf:l %.r and the degree of xp is Ipi = Lf:l Pi?
The kernel
can then be described by features wp(x) = JCiPTxp since k(x,y) = Lp wp(x)wp(y),
where the summation runs over all multi-indices of degree up to K. To assure that
the features are real, we assume that the coefficients Ck in the kernel are nonnegative. A hyperplane in feature space is parameterized by a weight vector w with
components w p, and if 0 < TI'W . W(xl'), a point (xl', TI') of the training set lies on
the correct side of the plane. To express that the plane P has maximal distance to
the points of the training set, we choose an arbitrary positive stability parameter /'i,
and require that the weight vector w* of P minimize w . w subject to the constraints
/'i, < TI'W' w(xl'), for f.l = 1, ... ,m.
2.1
The Statistical Mechanics Formulation
Statistical Mechanics allows to analyze a variety of learning scenarios exactly in the
"thermodynamic limit" of high input dimensionality, when the data distributions
are simple enough. In this approach, one computes a partition function which
serves as a generating function for important averages such as the generalization
error. To define the partition problem for SVMs one first analyzes a soft version of
the optimization problem characterized by an inverse temperature f3. One considers
the partition function
z=
f dwe- ~f3w.w IT
8(TI'W'
w(xl') -
/'i,),
(1)
1'=1
where the SVM constraints are enforced strictly using the Heaviside step function
8. Properties of w * can be computed from In Z and taking the limit f3 -t 00.
To model the training data, we assume that the random and independent input
components have zero mean and variance liN . This scaling assures that the variance of w . w(xl') stays finite in the large N limit. For the target rule we assume
that its deterministic part is given by the polynomial t(x) = Lp JCiPTBpxp with
real parameters Bp. The magnitude of the contribution of each degree k to the
value of t(x) is measured by the quantities
1
Tk = Ck Nk
'"'
~ Bp2
(2)
p,lpl =k
where Nk = (N+; - I) is the number of terms in the sum. The degree of t(x) is L
and lower than K, so TL > 0 and TL+l = ... = TK = O. Note, that this definition
of t(x) ensures that any feature necessary for computing t(x) is available to the
SVM. For brevity we assume that the constant term in t(x) vanishes (To = 0) and
the normalization is Lk Tk = 1.
2.2
The Noise Model
In the deterministic case the label of a point x would simply be the sign of t(x).
Here we consider a nondeterministic rule and the output label is obtained using a
random variable Tu E {-1, 1} parameterized by a scalar u. The observable instances
of the rule, and in particular the elements of the training set, are then obtained by
independently sampling the random variable (x, Tt(x))' Simple examples are additive
noise, Tt(x) = sgn(t(x) + 77), or multiplicative noise, Tt(x) = sgn(t(x)77), where 77 is
a noise term independent of x. In general, we will assume that the noise does not
systematically corrupt the deterministic component, formally
1
(3)
1> Prob(Tu = sgn(u)) >"2 for all u.
So sgn( t( x)) is the best possible prediction of the output label of x, and the minimal
achievable generalization error is fmin = (8( -t(X)Tt(x)))x. In the limit of many
input dimensions N, a central limit argument yields that for a typical target rule
fmin = 2(8( -u)0(u))u , where u is zero mean and unit variance Gaussian. The
function 0 will play a considerable role in the sequel. It is a symmetrized form of
the probability p(u) that Tu is equal to 1, 0(u) = ~(p(u) + 1 - p( -u)).
2.3
Order Parameter Equations
One now evaluates the average of In Z (Eq. 1) over random drawings of training
data for large N in terms of t he order parameters
Q
r
(((W.1]i(X))2)Jw'
q=((w)w?1]i(X))2)x
and
Q-! ?w ?1]i(x))w B ?1]i(x))x .
(4)
Here the thermal average over w refers to the Gibbs distribution (1). For the large
N limit, a scaling of the training set size m must be specified, for which we make
t he generic Ansatz m = aNt, where I = 1, ... ,L. Focusing on the limit of large j3,
where the density on the weight vectors converges to a delta peak and q -+ Q, we
introduce the rescaled order parameter X = j3( Q - q) / St, with
t
St = i (1) -
L
Ci .
(5)
i=O
Note that this scaling with St is only possible since the degree K of the kernel
i(x, y) is greater than I, and thus St ?- O. Finally, we obtain an expression for it =
lim,B--+oo limN --+00 ?In Z)) St / (j3Nt ), where the double brackets denote averaging over
all training sets of size m . The value of it results from extremizing, with respect to
r, q and X, the function
it(r,q,X) =
-aq /0(-u)G (ru
X
\
+ ~v -~))
v0 u,v
~ (~: - X ~ 1) (1- -(X -1)TzS~;Ct + L~=l TJ
(6)
where G(z) = 8(z)z2, and u, v are independent Gaussian random variables with
zero mean and unit variance.
?
Since the stationary value of it is finite , w* . w*)) is of the order Nt. So the
higher order components of w* are small, (W;)2 ? 1 for Ipi > I, although these
components playa crucial role in ensuring that a hyperplane separating the training
points exists even for large a. But the key quantity obtained from Eq. (6) is the
stationary value of r which determines the generalization error of the SVM via
fg = (0(-u)8(ru + ~v))u,v, and in particular fg = fmin for r = 1.
2.4
Universal Asymptotics
We now specialize to the case that l equals L, the degree of the polynomial t(x) in
the target rule. So m = aNL and for large a, after some algebra, Eq. (6) yields
r = 1where
B(q)
A(q*) ~
4B(q*)2 a
(e(Y)8(-Y+Ii/yrj))}y
(e(Y)8(-Y+Ii/y7i) (_Y+Ii/y7i)2}y.
(7)
and
A(q)
Further q* = argminqqA(q), and con-
sidering the derivatives of qA(q) for q --+ 0 and q --+
condition (3) assures that qA(q) does have a minimum.
00,
one may show that
Equation (7) shows that optimal generalization performance is achieved on this scale
in the limit of large a. Remarkably, as long as K > L, the asymptote is invariant
to the choice of the kernel since A(q) and B(q) are defined solely in terms of the
target rule.
3
Extension to other Kernels
Our next goal is to understand cases where the kernel is a general function of the
inner product or of the distance between the vectors. We still assume that the
target rule is of finite complexity, i.e. defined by a polynomial and corrupted by
noise and that the number of examples is polynomial in N. Remarkably, the more
general kernels then reduce to the polynomial case in the thermodynamic limit.
Since it is difficult to find a description of the Hilbert space for k( x, y) which is useful
for a Statistical Physics calculation, our starting point is the dual representation:
The weight vector w* defining the maximal margin hyperplane P can be written
as a linear combination of the feature vectors w(x M ) and hence w* . w(y) = IJ(Y),
where
m
(8)
M=l
By standard results of convex optimization theory the AM are uniquely defined by the
Kuhn-Tucker conditions AM ::::: 0, TMIJ(X M) ::::: Ii (for all patterns), further requiring
that for positive AM the second of the two inequalities holds as an equality. One also
finds that w* . w* = 2:;=1 AM and for a polynomial kernel we thus obtain a bound
on 2:;=1AM since w* . w* is O(m).
We first consider kernels ?(x? y), with a general continuous function ? of the inner
product, and assume that ? can be approximated by a polynomial J in the sense
that ?(1) = J(l) and ?(z) - J(z) = O(ZK) for z --+ O. Now, with a probability
approaching 1 with increasing N, the magnitude of xM?xl/ is smaller than, say, N- 1/3
for all different indices {t and v as long as m is polynomial in N. So, considering Eq.
(8), for large N the functions ?(z) and J(z) will only be evaluated in a small region
around zero and at z = 1 when used as kernels of a SVM trained on m = aNL
examples. Using the fact that 2:;=1AM = O(m) we conclude that for large Nand
K > 3L the solution of the K uhn-Tucker conditions for J converges to the one for
?. So Eqs. (6,7) can be used to calculate the generalization error for ? by setting
ttl = ?(l) (O)/l! for l = 1, ... , L, when ? is an analytic function. Note that results in
[4] assure that ttl ::::: 0 if the kernel ?( X? y) is positive definite for all input dimensions
N. Further, the same reduction to the polynomial case will hold in many instances
where ? is not analytical but just sufficiently smooth close to O.
3.1
RBF Kernels
We next turn to radial basis function (RBF) kernels where k( x, y) depends only
on the Euclidean distance between two inputs, k(x,y) = <I>(lx - YI2). For binary
input components (Xi = ?N- 1 / 2 ) this is just the inner product case since <I>(lx Y12) = <I>(2 - 2x? y). However, for more general input distributions, e.g. Gaussian
input components, the fluctuations of Ixl around its mean value 1 have the same
magnitude as x . y even for large N, and an equivalence with inner product kernels
is not evident.
Our starting point is the observation that any kernel <I>(lx - Y12) which is positive
definite for all input dimensions N is a positive mixture of Gaussians [6]. More
precisely <I>(z) = fooo e-ez da(k) where the transform a(k) is nondecreasing. For
the special case of a single Gaussian one easily obtains features 'IT p by rewriting
<I>(lx - Y12) = e-lx-v I2/2 = e- 1x12/2ex've-lvI2 /2. Expanding the kernel e X ' v into
polynomial features, yields the features 'IT p(x) = e- 1x12 /2x pl
for <I>(lx _ YI2).
But, for a generic weight vector w in feature space,
M
w? 'IT(x) = ~Wp'ITp(x) = e-~lxI2 ~wp
M
(9)
is of order 1, and thus for large N the fluctuations of Ixl can be neglected.
This line of argument can be extended to the case that the kernel is a finite mixture
of Gaussians, <I>(z) = L~=l aie-'Y7z /2 with positive coefficients ai. Applying the
reasoning for a single Gaussian to each term in the sum, one obtains a doubly
indexed feature vector with components 'lTp,i(X) = e-'Y7IxI2/2(ai/';lpl/lpll)1/2xp. It
is then straightforward to adapt the calculation of the partition function (Eq. 16) to the doubly indexed features, showing that the kernel <I>(lx - Y12) yields the
same generalization behavior as the inner product kernel <I> (2 - 2x . y). Based on the
calculation, we expect the same equivalence to hold for general radial basis function
kernels, i.e. an infinite mixture of Gaussians, even if it would be involved to prove
that the limit of many Gaussians commutes with the large N limit.
4
Simulations
To illustrate the general results we first consider a scenario where a linear target rule,
corrupted by additive Gaussian noise, is learned using different transcendental RBF
kernels (Fig. 1) . While Eq. (7) predicts that the asymptote of the generalization
error does not depend on the kernel, remarkably, the dependence on the kernel
is very weak for all values of a. In contrast, the generalization error depends
substantially on the nature of the noise process. Figure 2 shows the finding for
a quadratic rule with additive noise for the cases that the noise is Gaussian and
binary. In the Gaussian case a 1/a decay of Eg to Emin is found, whereas for binary
noise the decay is exponential in a. Note that in both cases the order parameter r
approaches 1 as 1/a.
5
Summary
The general characterization of learning curves obtained in this paper demonstrates
that support vector machines with high order or even transcendental kernels have
definitive advantages when the training data is noisy. Further the calculations leading to Eq. (6) show that maximizing the margin is an essential ingredient of the
approach: If one just chooses any hyperplane which classifies the training data correctly, the scale of the learning curve is not determined by the target rule and far
more examples are needed to achieve good generalization. Nevertheless our findings
are at odds with many of the current t heoretical motivations for maximizing the
margin which argue that this minimizes the effective Vapnik Chervonenkis dimension of the classifier and thus ensures fast convergence of the training error to the
generalization error [1 , 2]. Since we have considered hard margins, in contrast to t he
generalization error, the training error is zero, and we find no convergence between
the two quantities. But close to optimal generalization is achieved since maximizing
the margin biases the SVM to explain as much as possible of the data in terms of a
low order polynomial. While the Statistical Physics approach used in this paper is
only exactly valid in the thermodynamic limit, the numerical simulations indicate
that the theory is already a good approximation for a realistic number of input
dimensions.
We thank Rainer Dietrich for useful discussion and for giving us his code for the simulations. The work of M.O. was supported by the EPSRC (grant no. GR/M81601)
and the British Council (ARC project 1037); R.U. was supported by the DFG and
the DAAD.
References
[1] C. Cortes and V. Vapnik. , Machine Learning, 20:273-297, 1995.
[2] N. Cristianini and J . Shawe-Taylor. Support Vector Machines. Cambridge U niversity Press , 2000.
[3] M. Opper and R. Urbanczik. Phys. Rev. Lett., 86:4410- 4413, 200l.
[4] R. Dietrich, M. Opper, and H. Sompolinsky. Phys. Rev. Lett., 82:2975 - 2978,
1999.
[5] S. Risau-Gusman and M. Gordon. Phys. Rev. E, 62:7092- 7099,2000.
[6] I. Schoenberg. Anal. Math, 39:811-841, 1938.
0.3
,-----r -- - - - - - - - - - - - - - - - - ,
(A)
(8)
(C)
(D)
0.2
D
6.
<>
0
(E)
tOg
0.1
- trllin -
o
-
-
-
-
-
10
5
20
15
a=P/N
Figure 1: Linear target rule corrupted by additive Gaussian noise rJ ((rJ) = 0, \rJ 2 ) =
1/16) and learned using different kernels. The curves are the theoretical prediction;
symbols show simulation results for N = 600 with Gaussian inputs and error bars
are approximately the size of the symbols. (A) Gaussian kernel, <I>(z) = e- kz with
k = 2/3. (B) Wiener kernel given by the non analytic function <I>(z) = e - e..jZ. We
chose c ~ 0.065 so that the theoretical prediction for this case coincides with (A).
(C) Gaussian kernel with k = 1/20, the influence of the parameter change on t he
learning curve is minimal. (D) Perceptron, ?(z) = z . Above a e ~ 7.5 vanishing
training error cannot be achieved and the SVM is undefined. (E) Kernel invariant
asymptote for (A,B,C).
0.1
-
-E~in-
-
o w-______
o
2
-
-
-
-
-
-
-
-
-
-
-
~______~_ _ _ _ _~_____ _ w
4
6
8
a = P/N2
Figure 2: A noisy quadratic rule (Tl = 0, T2 = 1) learned using the Gaussian kernel
with k = 1/20. The upper curve (simulations.) is for additive Gaussian noise as
in Fig. 1. The lower curve (simulations .) is for binary noise, rJ ? s with equal
probability. We chose s ~ 0.20 so that the value of fmin is the same for the two
noise processes. The inset shows that f9 decays as l/a for Gaussian noise, whereas
an exponential decay is found in the binary case. The dashed curves are the kernel
invariant asymptotes. The simulations are for N = 90 with Gaussian inputs, and
standard errors are approximately the size of the symbols.
| 2123 |@word mild:2 version:1 achievable:1 polynomial:15 physik:2 simulation:7 commute:1 reduction:1 itp:1 chervonenkis:1 denoting:1 current:1 z2:1 nt:1 universality:1 intriguing:1 must:1 written:1 transcendental:2 realistic:3 partition:4 additive:5 numerical:1 analytic:2 asymptote:5 stationary:2 plane:3 vanishing:1 characterization:2 math:1 lx:7 ipi:2 become:1 specialize:1 doubly:2 prove:1 fitting:2 nondeterministic:1 introduce:1 manner:1 behavior:1 mechanic:3 multi:2 considering:2 increasing:1 becomes:1 begin:1 classifies:1 moreover:2 project:1 ttl:2 substantially:2 minimizes:1 finding:3 ti:4 exactly:2 demonstrates:1 classifier:1 uk:2 unit:2 grant:1 arguably:1 positive:6 engineering:1 limit:13 solely:1 fluctuation:2 approximately:2 might:1 chose:2 equivalence:2 suggests:1 practical:2 implement:2 lf:2 definite:2 urbanczik:2 asymptotics:2 universal:3 sidering:1 radial:2 refers:1 suggest:1 cannot:1 close:2 applying:1 influence:1 deterministic:3 maximizing:3 straightforward:1 starting:2 independently:1 convex:2 rule:20 his:1 stability:1 schoenberg:1 target:16 play:1 us:1 element:2 assure:2 approximated:1 bp2:1 predicts:1 role:3 epsrc:1 solved:1 calculate:2 region:1 ensures:2 sompolinsky:1 rescaled:1 vanishes:1 complexity:7 ock:1 cristianini:1 neglected:1 lvi2:1 trained:1 depend:2 raise:1 algebra:1 tog:1 basis:2 easily:1 fast:1 effective:1 widely:1 say:1 drawing:1 ability:3 transform:1 noisy:6 nondecreasing:1 advantage:3 dietrich:2 analytical:1 product:8 maximal:2 tu:3 rapidly:1 flexibility:2 achieve:2 opperm:1 lpl:2 description:1 inputoutput:1 convergence:3 regularity:2 double:1 oscillating:1 generating:1 converges:2 tk:3 oo:1 illustrate:1 ac:1 measured:1 ij:1 school:1 eq:8 indicate:1 kuhn:1 drawback:1 correct:2 stochastic:1 sgn:4 require:2 generalization:20 summation:1 strictly:1 extension:1 pl:1 hold:4 sufficiently:2 around:2 considered:1 mapping:3 achieves:2 birmingham:1 label:3 council:1 individually:1 gaussian:16 rather:3 ck:2 pn:1 rainer:1 fur:1 contrast:3 am:7 sense:1 helpful:1 inference:1 typically:1 nand:1 wurzburg:2 relation:1 germany:1 classification:2 dual:1 special:1 equal:4 construct:1 f3:2 having:1 sampling:1 t2:1 gordon:1 ve:1 dfg:1 investigate:1 mixture:3 bracket:1 undefined:1 tj:1 necessary:2 institut:1 indexed:2 euclidean:2 taylor:1 desired:1 theoretical:3 minimal:2 instance:2 classify:2 soft:1 introducing:1 gr:1 corrupted:4 chooses:1 st:5 density:1 peak:1 stay:1 sequel:1 physic:5 ansatz:1 central:1 choose:2 m81601:1 derivative:1 leading:1 de:1 coefficient:2 depends:2 multiplicative:1 analyze:1 characterizes:1 complicated:1 contribution:1 minimize:1 wiener:1 variance:4 largely:1 efficiently:1 yield:4 theoretische:1 ant:1 weak:1 daad:1 basically:1 published:1 explain:1 phys:3 definition:1 evaluates:1 tucker:2 involved:1 elegance:1 naturally:1 con:1 lim:1 dimensionality:1 hilbert:3 carefully:1 focusing:1 higher:1 emin:1 jw:1 formulation:1 evaluated:2 just:3 nonlinear:1 defines:1 perhaps:1 fooo:1 contain:1 requiring:1 hence:1 equality:1 y12:4 wp:5 i2:1 eg:1 uniquely:1 coincides:1 evident:1 tt:4 temperature:1 reasoning:1 image:1 extremizing:1 novel:1 common:2 b4:1 he:4 gusman:1 cambridge:1 gibbs:1 ai:2 shawe:1 aq:1 longer:1 v0:1 playa:1 recent:1 rence:1 scenario:5 inequality:1 binary:6 yi:1 analyzes:1 greater:1 minimum:1 converge:1 dashed:1 ii:4 full:1 thermodynamic:3 rj:4 smooth:2 match:2 characterized:1 adapt:1 calculation:4 long:2 lin:1 ensuring:1 prediction:4 j3:2 kernel:45 normalization:1 achieved:3 justified:1 whereas:2 remarkably:4 limn:1 crucial:1 extra:1 subject:1 ltp:1 odds:1 enough:1 variety:1 architecture:1 perfectly:1 approaching:1 inner:8 reduce:1 whether:1 expression:1 useful:2 amount:1 svms:8 exist:2 sign:2 delta:1 correctly:1 express:1 group:1 key:1 nevertheless:1 rewriting:1 utilize:1 asymptotically:1 sum:2 wand:1 run:1 inverse:1 parameterized:2 powerful:1 enforced:1 prob:1 decide:1 scaling:3 bound:1 ct:1 pay:1 quadratic:2 nonnegative:1 dangerous:1 constraint:2 precisely:1 bp:1 fmin:4 argument:3 x12:2 combination:2 smaller:1 lp:2 rev:3 modification:1 invariant:3 computationally:1 equation:2 previously:1 assures:2 turn:1 needed:1 serf:1 generalizes:1 available:1 gaussians:4 generic:2 symmetrized:1 slower:1 original:1 assumes:1 giving:2 already:2 question:1 quantity:3 dependence:1 dwe:1 guessing:1 distance:4 thank:1 separating:2 argue:1 considers:3 assuming:1 ru:2 code:1 index:3 difficult:1 robert:1 negative:1 aie:1 anal:1 reliably:1 proper:1 upper:1 observation:1 arc:1 finite:4 thermal:1 situation:1 defining:1 extended:1 arbitrary:2 specified:1 learned:3 qa:2 suggested:1 bar:1 pattern:1 xm:1 belief:1 aston:2 picture:1 lk:1 asymptotic:3 expect:2 ixl:2 ingredient:1 spoil:1 degree:9 xp:3 systematically:1 corrupt:1 pi:3 uhn:1 summary:1 surprisingly:1 supported:2 side:2 allow:1 understand:1 bias:1 perceptron:1 taking:1 fg:2 curve:10 dimension:6 opper:2 valid:1 lett:2 computes:1 kz:1 far:2 observable:1 uni:1 obtains:2 assumed:1 conclude:1 anl:2 xi:1 continuous:1 learn:1 zk:2 nature:1 expanding:1 jz:1 complex:1 da:1 main:1 yi2:2 motivation:1 noise:19 definitive:1 n2:1 fig:2 f9:1 tl:3 exponential:2 xl:9 lie:2 ix:1 british:1 inset:1 showing:1 symbol:3 decay:7 svm:13 cortes:1 exists:1 essential:1 vapnik:2 ci:1 magnitude:3 margin:6 nk:2 suited:1 simply:1 ez:1 expressed:1 scalar:1 determines:1 goal:1 rbf:3 price:1 universitiit:1 feasible:1 analysing:1 considerable:1 hard:1 infinite:3 determined:2 typical:1 change:1 hyperplane:6 averaging:1 risau:1 formally:1 wuerzburg:1 support:5 latter:1 brevity:1 yrj:1 evaluate:1 heaviside:1 ex:1 |
1,234 | 2,124 | Circuits for VLSI Implementation of
Temporally-Asymmetric Hebbian
Learning
Adria Bofill
Alan F. Murray
DanlOn P. Thompson
Dept. of Electrical Engineering
The University of Edinburgh
Edinburgh , EH93JL , UK
adria. [email protected]. uk
alan. murray @ee.ed.ac.uk
damon. thompson @ee.ed.ac. uk
Abstract
Experimental data has shown that synaptic strength modification
in some types of biological neurons depends upon precise spike timing differences between presynaptic and postsynaptic spikes. Several temporally-asymmetric Hebbian learning rules motivated by
this data have been proposed. We argue that such learning rules
are suitable to analog VLSI implementation. We describe an easily tunable circuit to modify the weight of a silicon spiking neuron
according to those learning rules. Test results from the fabrication
of the circuit using a O.6J.lm CMOS process are given.
1
Introduction
Hebbian learning rules modify weights of synapses according to correlations between
activity at the input and the output of neurons. Most artificial neural networks
using Hebbian learning are based on pulse-rate correlations between continuousvalued signals; they reduce the neural spike trains to mean firing rates and thus
precise timing does not carry information. With this approach the spiking nature
of biological neurons is just an efficient solution that evolution has produced to
transmit analog information over an unreliable medium.
In recent years, recorded data have indicated that synaptic strength modifications
are also induced by timing differences between pairs of presynaptic and postsynaptic
spikes [1][2]. A class of learning rules derived from these experimental dat a is illustrated in Figure 1 [2]-[4]. The "causal/non-causal" basis of these Hebbian learning
algorithms is present in all variants of this spike-timing dependent weight modification rule. When the presynaptic spike arrives at the synapse a few milliseconds
presynaptic spike
presynaptic spike
,tpre
tpre'
postsynaptic spike
postsynaptic spike
!post
tpost '
!'.w
tpre - tpost
tpre - tpost
(a)
(b)
Figure 1: Two temporally-asymmetric Hebbian learning rules drawing on
experimental data. The curves show the shape of the weight change (~W) for
differences between the firing times of the presynaptic (tpr e) and the postsynaptic
(tpost) neurons. When the presynaptic spike arrives at the synapse a few ms before the postsynaptic neuron fires , the weight of the synapse is increased. If the
postsynaptic neuron fires first, the weight is decreased.
before an output spike is generated, the synaptic efficiency increases. In contrast,
when the postsynaptic neuron fires first , the efficiency of the synapse is weakened.
Hence, only those synapses that receive spikes that appear to contribute to the
generation of the postsynaptic spike are reinforced. In [5] a similar spike-timing
difference based learning rule has been used to learn input sequence prediction in a
recurrent network. Studies reported in [4] indicate that the positive (potentiation)
element of the learning curve must be smaller than the negative (depression) to
obtain stable competitive weight modification.
Pulse signal representation has been used extensively in hardware implementations
of artificial neural networks [6] [7]. Such systems use pulses as a mere technological
solution to benefit from the robustness of binary signal transmission while making
use of analog circuitry for the elementary computation units. However , they do not
exploit the relative timing differences between individual pulses to compute. Also ,
analog hardware is not well-suited to the complexity of most artificial neural network
algorithms. The learning rules presented in Figure 1 are suitable for analog VLSI
because: (a) the signals involved in the weight modification are local to the neuron ,
(b) no temporal averaging of the presynaptic or postsynaptic activity is needed and
(c) they are remarkably simple compared to complex neural algorithms that impose
mathematical constraints in terms of accuracy and precision. An analog VLSI
implementation of a similar, but more complex, spike-timing dependent learning
rule can be found in [8].
We describe a circuit that implements the spike-timing dependent weight change
described above along with the t est results from a fabricated chip. We have fo cused
on the implem entation of the weight modification circuits, as VLSI spiking neurons
with tunable m embrane time constant and refractory p eriod have already b een
proposed in [9] and [10].
2
Learning circuit description
Figure 2 shows the weight change circuit and Figure 3 the form of signals required
to drive learning. These driving signals are generated by the circuits described in
Figure 4. The voltage across the weight capacitor , Cw in Figure 2, is modified
according to t he spike-timing dependent weight change rule discussed above. The
weight change, ~ W, is defined as -~ Vw so that the leakage of t he capacitor leads
Vw in the direction of weight decay. The circuits presented allow the control of:
(a) the abruptness of the transition between potentiation and depression at the
origin, (b) the difference between the areas under the curve in the potentiation and
depression regions, (c) the absolute value of the area under each side of the curve
and (d) the time constant of t he curve decay.
PI
Figure 2: W e ight change circuit
postsynaptic spike
up
n '- - -__
down
(a)
(b)
Figure 3: Stimulus for the w e ight change circuit
The weight change circuit of Figure 2 works as follows. When a falling edge of
either a postsynaptic or a presynaptic spike occurs , a short activation pulse is
generated which causes Cd ec to be charged to V pea k through transistor Nl. The
charge accumulated in Cd ec will leak to ground with a rate set by Vd ec ay ' The
resulting voltage at the gate of N3 produces a current flowing through P2-P3-N4. If
a presynaptic spike is active after the falling edge of a postsynaptic spike an activelow up pulse is applied to the gate of transistor P5. Thus, the current flowing
through N3 is mirrored to transistor P4 causing an increase in the voltage across
Cw that corresponds to a decrease in the weight. In contrast, when a presynaptic
spike precedes a postsynaptic spike an active-high down pulse is generated and the
current in N3 is mirrored to N5-N6 resulting in a discharge of Cw .
As the current in N2 is constant, the current integrated by Cw displays an exponential decay, if Vpeak is such that N3 is in sub-threshold mode. Hence, the rate
of decay of the learning curve is fixed by the ratio hlCdec. The abruptness of the
transition zone between potentiation and depression is set by the duration of both
the presynaptic and postsynaptic spike. Finally, an imbalance between the areas
under the positive and negative side of the curve can be introduced via Vdep and
Vpot . The effect of all these circuit parameters is exemplified by the test results
shown in the following section.
act
down
post_spike
(a)
(b)
Figure 4: Learning drivers. (a) Delayed act pulse generator. (b) Asynchronous controller for up and down signals
The circuit of Figure 4(a) , present in both the presynaptic and postsynaptic neurons,
generates a short act pulse with the falling edge of the output spike. The act pulses
are ORed at each synapse to produce the activation pulse applied to the weight
change circuit of Figure 2.
The other two driving signals , up and down, are produced by a small asynchronous
controller using standard and asymmetric C-elements [11] shown in Figure 4(b).
The internal signal q indicates if the last falling edge to occur corresponds to a
pre (q = 1) or a postsynaptic spike (q = 0). This ensures that an up signal that
decreases the weight is only generated when a presynaptic spike is active after the
falling edge of a postsynaptic spike. Similarly down is activated only when the
postsynaptic spike is active following a presynaptic spike falling edge.
Using the current flowing through N3 (Figure 2) to both increase and decrease the
weight allows us to match the curve at the potentiation and depression regions at
the exp ense of having to introduce the driving circuits of Figure 4.
3
Results from the temporally-asymmetric Hebbian chip
The circuit in Figure 2 has b een fabricated in a O.6J.lm standard CMOS process.
The driving signals (down, up and activation) are currently generated off-chip.
The circuit can be operated in t he p,s timescale, however , here we only present test
results with time constants similar to those suggested by experimental data and
studied using software models in [3]- [5].
3.5"==;;;r----;;.==---~--~==~
Vdecay = 515mV
t
2.5
t
pre
-I
- t
Vpeak = 694m V
= 1ms
T
'p
T act = 50j.ts
=2ms
V
\
pre
post
pot
=OV
Vdd - Vdep = OV
post
=5
t
Vdecay = 515mV
post
>'
V peak = 694mV
= 1ms
-I pre =5m
/
1.5
t
T
'p
= SOj.tS
T act
t
pre
-I
po st
V
=7.5ms
pot
=OV
tpost - tpre
Vdd - Vdep = OV
0.5
0.5
1.5
= 7.5ms
ms
/
00
2.5
=
i
- t
po"
0.5
1.5
Time ( s )
Time(s)
(a)
2.5
(b)
Figure 5: Test result s . Linearity. (a ) The voltage across Cw is initially set to
OV and increased by a sequence of consecutive pairs of pre and postsynaptic spikes.
The delays between presynaptic and postsynaptic firing times were set to 2ms , 5ms
and 7.5ms (b) The order of pre and postsynaptic spikes is reversed to decrease Vw .
In both plots the duration of the spikes, T sp , and the activation pulse, Ta ct , is set
to 1ms and and 50p,s respectively.
The learning window plots shown in Figures 6-8 were constructed with test data
from a sequence of consecutive presynaptic and postsynaptic spikes with different
delays . Before every new pair of presynaptic and postsynaptic spikes, the voltage
in Cw was reset to Vw = 2V . The weight change curves are similar for other initial
"reset" weight voltages owing to the linearity of the learning circuit for different Vw
values as shown in Figure 5. A power supply voltage of Vdd = 5V is used in all test
results shown.
80 , - - . - - , - - , - - , - - , - - , - - , - - - ,
V decay =516mV
100
Tsp =1ms
,,'k
=
Tact 501-15
V
=OV
>
50
""
pot
Vdd - Vdep = OV
-
Tsp= 1ms
Vpeak =716mV
--- V
=711mV
-- V
=701mV
60
Tact = 5O~IS
40
Vdd - Vdep = OV
Vpot = OV
>
E
E 20
0
~
-50
- - - v decay = 499mV
"
""
, ",
Vpeak = 701mV
- -
- - . . . . .~--.=.::.---.:.-
>~
>'
<l
Vdecay =517mV
Vpeak = 702mV
- ';",-;"--' ::"= '- -"'- -
l'
r
Vdecay =482mV
Vpeak = 699mV
,I
,
; -20
<l
-40
-60
- 100
-80~-~-~-~-~-~-~-~-----'
-25 -20
-15 - 10
-5
t p ,.
0
-
5
tpo,t (ms)
(a)
10
15
20
25
-40
-30
-20
-10
0
10
20
30
40
t pre - \ OSI (m s)
(b)
Figure 6: Test result s . (a) M aximum w e ig ht ch ange . (b) Learning window
decay. The decay of both tails of the learning window is set by Vdecay. A wid e
range of time constants can b e set. Note, however, that Vpea k needs to b e increased
slightly for faster decay rates to maintain exactly the same p eak value.
The maximum weight change is easily tuned with Vp e ak as shown in Figure 6( a).
Changing the value of Vp e ak modifies by the same amount the absolute value of the
peaks at both sides of the curve. The decay of the learning window is controlled
by Vd ec ay' An increase in Vd ecay causes both tails of the learning window to decay
faster as seen in Figure 6 (b). As m entioned above, matching between both sides of
the learning window is possible because the same source of current is used to both
increase and decrease the weight.
100
80
-
>
~-~-~~~~-~-~-~-~
Vdecay = SOOmV
Vpeak = 705mV
60
= 1ms
'p
Tact =50l1s
40
Vpot =OV
T
80
=9.3mV
60
_ ,_
dd
dep
Vdd - Vdep
>
E
>'0
I'
-20
<l
-40
;:
Vdecay = SOOmV
Vpeak = 705mV
Tsp= 1ms
Tact =
.:. 20
-
100
Vdd - Vdep = OV
- - - V - V = 3.8mV
40
,,_
50~lS
pot
,, ; '
Vdd -Vdep=OV
20
V =OV
pot
- - - V =47mV
- -- VPOt = 9~V
//
.-."';::: ,-;//
>' 0
,
<l
;:
<l
-60
-20
-40
-60
-80
-80
-1~go'---_--c1~
5 ---C10~~-5~-0~-~5-~-~
15:--~
-100
-20
20
- 15
- 10
-5
'pre - 'post (ms)
0
5
10
15
20
'pre - 'post (ms)
(a)
(b)
Figure 7: Test results. Imbalance between potentiation and depression.
The imbalance between the areas under the potentiation and depression regions of
the learning window is a critical parameter of this class of learning rules [3] [4]. The
circuit proposed can adjust the peak of the curve for potentiation and depression
independently (Figure 7). Vp ot can be used to reduce the area under the potentiation
region while keeping unchanged the depression part of the curve , thus setting the
overall area under the curve to a negative value (Figure 7(b)). Similarly, with
Vdd - Vd ep the area of the depression region can also be reduced (Figure 7 (a) ) .
100 ,---,--~--,,---,--~--,
100
50
_ _ Tsp = 100llS
Vpeak = 790mV
Vdecay = 499mV
Tact = 50flS
Vdd - Vdep =OV
Vpot =OV
---- T
"
>
E
50
i
100
=1ms
Vpeak = 699mV
Vdecay = 482mV
;: 100
<l
50
50
100 '---~--~--L--~--~-~
-15
-10
-5
0
5
10
15
'pre - 'post (ms)
Figure 8: Test results. Abruptness at the origin
The abruptness of the learning window at the origin (short delays between pre and
postsynaptic spikes) is set by th e duration of the spikes. Dat a in Figure 8 show
that th e two peaks of the learning window are separat ed by 2 times the durations
of the spikes (Tsp ).
4
Discussion and future work
Drawn from experimental data, several temporally-asymmetric Hebbian learning
rules have been proposed recently. These learning rules only strengthen the weights
when there is a causal relation between presynaptic and postsynaptic activity.
Purely random time coincidences between spikes will tend to decrease the weights.
Synaptic weight normalization is thus achieved via competition to drive postsynaptic spikes [4]. Predictive sequence learning has been achieved using a similar
time-difference learning rule based on the same data [5]. Other pulse-based learning rules have also been used to study how delay tuning could be achieved in the
sound source localization system of the barn owl [12].
A simple circuit to implement a general weight change block based on such learning
rules has been designed and partially fabricated. The main characteristics of the
learning rule, namely the abruptness at the origin, the rate of the decay of the
learning window, the imbalance between the potentiation and depression regions
and the rate of learning , can be tuned easily. The design also ensures that the
circuit can operate at different timescales. As shown, the fabricated circuits have
good linearity over a wide range of weight voltage values.
We are currently developing a second chip with a small network of temporally
asymmetric Hebbian spiking neurons using the circuit described in this paper. The
structure of the network will be reconfigurable. The small network will be used to
carry out movement planning exp eriments by learning of temporal sequences. We
envisage the application of networks of temporally-asymmetric Hebbian learning
silicon neurons as higher level processing stages for the integration of sensor and
motor activities in neuromorphic system. We will concentrate on auditory applications and adaptive, spike-based motion estimation. In both types of application,
naturally-occurring correlations in data can b e exploited to drive the pulse timingbased learning process.
Acknowledgelllents
We thank Robin Woodburn , Patrice Fleury and Martin Reekie for fruitful discussions during the design and tape out of the chip. We also acknowledge that the
circuits presented incorporate some of the insights into neuromorphic engineering
that one of the authors gained at the Telluride Workshop on Neuromorphic Engineering 2000 (http://www.ini.unizh.ch/telluride2000 /).
References
[1] Markram, H., Lubke, J., Frotscher , M. & Sakmann, B.(1997) Regulation of Synaptic
Efficacy by Coincidence of Postsynaptic APs and EPSPs. Science 275 , 213-215.
[2] Zhang , L.L , Tao, H.W., Holt, C.E. , Harris, W.A. & Poo , M-m.(1998) A critical window
for cooperation and competition among developing retinotectal synapses. Nature 395 , 3744.
[3] Abbott, L.F. & Song, S.(1999) Temporally Asymmetric Hebbian Learning , Spike Timing and Neuronal Response Variability. In Kearns, M.S., Solla, S.A., & Cohn, D.A. (eds.),
Advances in Neura l Information Processing S ystems 11, 69-75. Cambridge, MA: MIT
Press.
[4] Song, S., Miller, K.D. & Abbo tt , L.F .(2000) Competitive Hebbian Learning Through
Spike-Timing Dependent Synaptic Plasticity. Nature Neuroscience 3, 919-926.
[5] Rao, R.P.N., & Sejnowski, T .J.(2000) Predictive Sequence Learning in Recurrent Neocortical Circuits. In Solla, S.A., Leen, T .K. & Muller, K-R. (eds.), Advances in Neural
Information Processing Systems 12, 164-170. Cambridge, MA: MIT Press.
[6] Murray, A.F. & Smith A.V.W.(1987) Asynchronous Arithmetic for VLSI Neural Syst ems. Electronic Letters 23 , 642-643.
[7] Murray, A.F. & Tarrasenko, L.(1994) Neural Computing: An Analogue VLSI Approach.
Chapman- Hall.
[8] Hafliger, P., Mahowald, M. & Watts , L. (1996) A Spike Based Learning Neuron in
Analog VLSL In Mozer , M.C., Jordan , M.L , & Petsche , T. (eds.) , Advances in Neural
Information Processing Systems 9, 692-698. Cambridge, MA: MIT Press.
[9] Indiveri, G.(2000) Modeling Selective Attention Using a Neuromorphic Analog VLSI
Device. Neural computation 12 , 2857-2 880.
[10] van Schaik, A., Fragniere, E. & Vittoz , E.( 1996) An Analogue Electronic Model of
Ventral Cochlear Nucleus Neurons. In Proceedings of the 5th International Conferen ce on
M icroelec tron ics for Neural, Fuzzy and Bio-inspired S ystems; Mi croneuro '96, 52-59. Los
Alamitos, CA: IEEE Computer Society Press.
[11] Shams, M., Ebergen, J.C. & Elmasry, M.L (1998) Modeling and Comparing CMOS
Implem entations of the C-elment. IEEE Tr ansactions on Very Large Scale Intergration
(VLSI) Systems , Vol. 6, No.4 , 563-567.
[12] Gerstner, W. , Kempter , R. , van Hemmen , J.1. & Wagner , H.(1999) Hebbian Learning
of Pulse Timing in the Barn Owl Auditory System. In Mass , W. & Bishop , C.M. (eds.) ,
Pulsed Neura l Networks. Cambridge , MA: MIT Press.
| 2124 |@word pulse:15 tr:1 carry:2 initial:1 efficacy:1 tuned:2 current:7 comparing:1 activation:4 must:1 plasticity:1 shape:1 motor:1 plot:2 designed:1 aps:1 device:1 smith:1 short:3 schaik:1 contribute:1 zhang:1 mathematical:1 along:1 constructed:1 driver:1 supply:1 introduce:1 planning:1 inspired:1 continuousvalued:1 window:11 linearity:3 circuit:26 medium:1 mass:1 fuzzy:1 aximum:1 fabricated:4 temporal:2 every:1 act:6 charge:1 exactly:1 uk:4 control:1 unit:1 bio:1 appear:1 before:3 positive:2 engineering:3 timing:12 modify:2 local:1 ak:2 firing:3 weakened:1 studied:1 range:2 block:1 implement:2 area:7 matching:1 acknowledgelllents:1 pre:12 holt:1 www:1 fruitful:1 charged:1 modifies:1 go:1 poo:1 attention:1 duration:4 thompson:2 l:1 independently:1 rule:18 insight:1 transmit:1 discharge:1 strengthen:1 origin:4 element:2 asymmetric:9 ep:1 p5:1 coincidence:2 electrical:1 vdep:9 region:6 ensures:2 solla:2 decrease:6 technological:1 movement:1 mozer:1 leak:1 complexity:1 vdd:10 ov:15 predictive:2 purely:1 upon:1 localization:1 efficiency:2 basis:1 tpo:1 easily:3 po:2 chip:5 train:1 describe:2 sejnowski:1 artificial:3 precedes:1 drawing:1 timescale:1 tsp:5 envisage:1 patrice:1 sequence:6 transistor:3 reset:2 p4:1 causing:1 osi:1 description:1 competition:2 los:1 transmission:1 produce:2 cmos:3 recurrent:2 ac:3 soj:1 dep:1 c10:1 p2:1 epsps:1 pot:5 indicate:1 vittoz:1 direction:1 concentrate:1 vpot:5 owing:1 pea:1 wid:1 owl:2 potentiation:10 biological:2 elementary:1 hall:1 ground:1 barn:2 exp:2 ic:1 fragniere:1 lm:2 circuitry:1 driving:4 ventral:1 consecutive:2 estimation:1 currently:2 mit:4 sensor:1 modified:1 voltage:8 derived:1 indiveri:1 indicates:1 contrast:2 dependent:5 accumulated:1 integrated:1 initially:1 vlsi:9 relation:1 selective:1 tao:1 overall:1 among:1 integration:1 frotscher:1 having:1 chapman:1 unizh:1 future:1 stimulus:1 lls:1 few:2 individual:1 delayed:1 fire:3 maintain:1 adjust:1 arrives:2 nl:1 operated:1 activated:1 separat:1 edge:6 causal:3 increased:3 modeling:2 rao:1 neuromorphic:4 mahowald:1 delay:4 fabrication:1 reported:1 st:1 peak:4 international:1 off:1 recorded:1 syst:1 mv:22 depends:1 competitive:2 abruptness:5 accuracy:1 lubke:1 characteristic:1 reinforced:1 miller:1 vp:3 produced:2 mere:1 drive:3 synapsis:3 fo:1 ed:8 synaptic:6 involved:1 naturally:1 mi:1 auditory:2 tunable:2 ta:1 higher:1 flowing:3 response:1 synapse:5 leen:1 ystems:2 just:1 stage:1 correlation:3 cohn:1 mode:1 indicated:1 effect:1 evolution:1 hence:2 illustrated:1 embrane:1 during:1 m:20 ini:1 ay:2 tt:1 neocortical:1 tron:1 motion:1 hafliger:1 recently:1 spiking:4 refractory:1 analog:8 he:4 tpr:1 discussed:1 tail:2 silicon:2 cambridge:4 tuning:1 vdecay:9 similarly:2 vlsl:1 stable:1 damon:1 recent:1 pulsed:1 binary:1 exploited:1 muller:1 seen:1 impose:1 ight:2 signal:11 arithmetic:1 sound:1 sham:1 fleury:1 hebbian:13 alan:2 match:1 faster:2 post:7 controlled:1 prediction:1 variant:1 n5:1 controller:2 normalization:1 achieved:3 c1:1 receive:1 remarkably:1 tact:5 decreased:1 source:2 ot:1 operate:1 induced:1 tend:1 capacitor:2 jordan:1 ee:3 vw:5 een:2 reduce:2 implem:2 motivated:1 song:2 cause:2 tape:1 depression:11 amount:1 extensively:1 hardware:2 reduced:1 http:1 mirrored:2 millisecond:1 neuroscience:1 vol:1 threshold:1 falling:6 drawn:1 changing:1 ce:1 abbott:1 ht:1 year:1 letter:1 electronic:2 p3:1 fl:1 ct:1 display:1 activity:4 strength:2 occur:1 constraint:1 n3:5 software:1 generates:1 martin:1 developing:2 according:3 watt:1 smaller:1 across:3 slightly:1 postsynaptic:28 em:1 n4:1 modification:6 making:1 needed:1 petsche:1 robustness:1 gate:2 eriments:1 exploit:1 neura:2 murray:4 society:1 dat:2 leakage:1 unchanged:1 already:1 alamitos:1 spike:45 occurs:1 cw:6 reversed:1 thank:1 vd:4 cochlear:1 presynaptic:19 argue:1 bofill:2 ratio:1 regulation:1 negative:3 implementation:4 design:2 sakmann:1 imbalance:4 neuron:15 acknowledge:1 t:2 variability:1 precise:2 introduced:1 pair:3 required:1 namely:1 tpre:5 suggested:1 exemplified:1 analogue:2 power:1 suitable:2 critical:2 temporally:8 n6:1 l1s:1 relative:1 kempter:1 generation:1 generator:1 nucleus:1 reekie:1 dd:1 pi:1 cd:2 cooperation:1 last:1 asynchronous:3 keeping:1 side:4 allow:1 wide:1 markram:1 wagner:1 absolute:2 edinburgh:2 benefit:1 curve:13 van:2 transition:2 author:1 tpost:5 adaptive:1 ig:1 ec:4 eak:1 unreliable:1 active:4 robin:1 retinotectal:1 nature:3 learn:1 ca:1 gerstner:1 complex:2 sp:1 main:1 timescales:1 n2:1 neuronal:1 hemmen:1 precision:1 sub:1 exponential:1 elmasry:1 down:7 bishop:1 reconfigurable:1 decay:12 workshop:1 gained:1 occurring:1 suited:1 partially:1 ch:2 corresponds:2 harris:1 ma:4 adria:2 telluride:1 change:12 averaging:1 kearns:1 experimental:5 est:1 zone:1 internal:1 incorporate:1 dept:1 |
1,235 | 2,125 | The Concave-Convex Procedure (CCCP)
A. L. Yuille and Anand Rangarajan *
Smith-Kettlewell Eye Research Institute,
2318 Fillmore Street,
San Francisco, CA 94115, USA.
Tel. (415) 345-2144. Fax. (415) 345-8455.
Email [email protected]
* Prof. Anand Rangarajan. Dept. of CISE, Univ. of Florida Room 301, CSE
Building Gainesville, FL 32611-6120 Phone: (352) 392 1507 Fax: (352) 392 1220
e-mail: [email protected]
Abstract
We introduce the Concave-Convex procedure (CCCP) which constructs discrete time iterative dynamical systems which are guaranteed to monotonically decrease global optimization/energy functions. It can be applied to (almost) any optimization problem and
many existing algorithms can be interpreted in terms of CCCP. In
particular, we prove relationships to some applications of Legendre
transform techniques. We then illustrate CCCP by applications to
Potts models, linear assignment, EM algorithms, and Generalized
Iterative Scaling (GIS). CCCP can be used both as a new way to
understand existing optimization algorithms and as a procedure for
generating new algorithms.
1
Introduction
There is a lot of interest in designing discrete time dynamical systems for inference
and learning (see, for example, [10], [3], [7], [13]).
This paper describes a simple geometrical Concave-Convex procedure (CCCP) for
constructing discrete time dynamical systems which can be guaranteed to decrease
almost any global optimization/energy function (see technical conditions in section (2)).
We prove that there is a relationship between CCCP and optimization techniques
based on introducing auxiliary variables using Legendre transforms. We distinguish
between Legendre min-max and Legendre minimization. In the former, see [6], the
introduction of auxiliary variables converts the problem to a min-max problem
where the goal is to find a saddle point. By contrast, in Legendre minimization, see
[8], the problem remains a minimization one (and so it becomes easier to analyze
convergence). CCCP relates to Legendre minimization only and gives a geometrical
perspective which complements the algebraic manipulations presented in [8].
CCCP can be used both as a new way to understand existing optimization algorithms and as a procedure for generating new algorithms. We illustrate this by
giving examples from Potts models, EM, linear assignment, and Generalized Iterative Scaling. Recently, CCCP has also been used to construct algorithms to
minimize the Bethe/Kikuchi free energy [13].
We introduce CCCP in section (2) and relate it to Legendre transforms in section (3). Then we give examples in section (4).
2
The Concave-Convex Procedure (CCCP)
The key results of CCCP are summarized by Theorems 1,2, and 3.
Theorem 1 shows that any function , subject to weak conditions, can be expressed
as the sum of a convex and concave part (this decomposition is not unique). This
implies that CCCP can be applied to (almost) any optimization problem.
Theorem 1. Let E(x) be an energy function with bounded Hessian [J2 E(x)/8x8x.
Then we can always decompose it into the sum of a convex function and a concave
function.
Proof. Select any convex function F(x) with positive definite Hessian with eigenvalues bounded below by f > o. Then there exists a positive constant A such that
the Hessian of E(x) + AF(x) is positive definite and hence E(x) + AF(x) is convex. Hence we can express E(x) as the sum of a convex part, E(x) + AF(x) , and a
concave part -AF(x).
Figure 1: Decomposing a function into convex and concave parts. The original function (Left Panel) can be expressed as the sum of a convex function (Centre Panel)
and a concave function (Right Panel). (Figure courtesy of James M. Coughlan).
Our main result is given by Theorem 2 which defines the CCCP procedure and
proves that it converges to a minimum or saddle point of the energy.
Theorem 2 . Consider an energy function E(x) (bounded below) of form E(x) =
Evex (x) + E cave (x) where Evex (x), E cave (x) are convex and concave functions of x
respectively. Then the discrete iterative CCCP algorithm ;zt f-7 ;zt+1 given by:
\1Evex
(x-t+l ) _- -\1Ecave
(x-t ),
(1)
is guaranteed to monotonically decrease the energy E(x) as a function of time and
hence to converge to a minimum or saddle point of E(x).
Proof. The convexity and concavity of Evex (.) and Ecave (.) means that Evex (X2) 2:
Evex (xd + (X2 -xd? ~ Evex (xd and Ecave (X4) :S Ecave (X3) + (X4 -X3)? ~ Ecave (X3 ),
for all X1 ,X2,X3,X4. Now set Xl = xt+l,X2 = xt,X3 = xt,X4 = xt+1. Using the
algorithm definition (i.e. ~Evex (xt+1) = -~Ecave (xt)) we find that Evex (xt+ 1) +
Ecave (xt+1) :S Evex (xt) + Ecave (xt), which proves the claim.
We can get a graphical illustration of this algorithm by the reformulation shown in
figure (2) (suggested by James M. Coughlan). Think of decomposing the energy
function E(x) into E 1(x) - E 2(x) where both E 1(x) and E 2(x) are convex. (This
is equivalent to decomposing E(x) into a a convex term E 1(x) plus a concave term
-E2(X)) . The algorithm proceeds by matching points on the two terms which have
the same tangents. For an input Xo we calculate the gradient ~ E2 (xo) and find the
point Xl such that ~ E1 (xd = ~ E2 (xo). We next determine the point X2 such that
~E1(X2) = ~E2 (X1)' and repeat.
7~------~--------~------,
60 50 40 -
30 -
20 -
o
10
O L---~=-~O-=~~~~----~
10
XO
Figure 2: A CCCP algorithm illustrated for Convex minus Convex. We want to
minimize the function in the Left Panel. We decompose it (Right Panel) into
a convex part (top curve) minus a convex term (bottom curve). The algorithm
iterates by matching points on the two curves which have the same tangent vectors,
see text for more details. The algorithm rapidly converges to the solution at x = 5.0.
We can extend Theorem 2 to allow for linear constraints on the variables X, for
example Li et Xi = aM where the {en, {aM} are constants. This follows directly
because properties such as convexity and concavity are preserved when linear constraints are imposed. We can change to new coordinates defined on the hyperplane
defined by the linear constraints. Then we apply Theorem 1 in this coordinate
system.
Observe that Theorem 2 defines the update as an implicit function of xt+ 1. In many
cases, as we will show, it is possible to solve for xt+1 directly. In other cases we may
need an algorithm, or inner loop , to determine xt+1 from ~Evex (xt+1). In these
cases we will need the following theorem where we re-express CCCP in terms of
minimizing a time sequence of convex update energy functions Et+1 (xt+1) to obtain
the updates xt+1 (i.e. at the tth iteration of CCCP we need to minimize the energy
Et+1 (xt+1 )). We include linear constraints in Theorem 3.
Theorem 3. Let E(x) = Evex (x) + E cave (x) where X is required to satisfy the linear
constraints Li et Xi = aM, where the {et}, {aM} are constants. Then the update rule
for xt+1 can be formulated as minimizing a time sequence of convex update energy
functions Et+1 (;rt+1):
(2)
where the lagrange parameters P'J1} impose linear comnstraints.
Proof. Direct calculation.
The convexity of EH1 (;rt+1) implies that there is a unique minimum corresponding
to ;rt+1. This means that if an inner loop is needed to calculate ;rt+1 then we can
use standard techniques such as conjugate gradient descent (or even CCCP).
3
Legendre Transformations
The Legendre transform can be used to reformulate optimization problems by introducing auxiliary variables [6]. The idea is that some of the formulations may
be more effective (and computationally cheaper) than others. We will concentrate
on Legendre minimization, see [7] and [8], instead of Legendre min-max emphasized
in [6]. An advantage of Legendre minimization is that mathematical convergence
proofs can be given. (For example, [8] proved convergence results for the algorithm
implemented in [7].)
In Theorem 4 we show that Legendre minimization algorithms are equivalent to
CCCP. The CCCP viewpoint emphasizes the geometry of the approach and complements the algebraic manipulations given in [8]. (Moreover, our results of the
previous section show the generality of CCCP while, by contrast, the Legendre
transform methods have been applied only on a case by case basis).
Definition 1. Let F(x) be a convex function. For each value y let F*(ff) =
minx{F(x) +y?x.}. Then F*(Y) is concave and is the Legendre transform of F(x).
Moreover, F (x) = max y { F* (y) - y. x} .
Property 1. F(.) and F*(.) are related by
a:; (fJ)
= {~~} - 1(_Y), -~~(x) =
{a{y* } -1 (x). (By { a{y* } -1 (x) we mean the value y such that a{y* (y) = x.)
Theorem 4. Let E1 (x) = f(x) + g(x) and E 2(x, Y) = f(x) + x? Y + h(i/), where
f(.), h(.) are convex functions and g(.) is concave. Then applying CCCP to E1 (x) is
equivalent to minimizing E2 (x, Y) with respect to x and y alternatively (for suitable
choices of g(.) and h(.).
Proof. We can write E1(X) = f(x) +miny{g*(Y) +x?y} where g*(.) is the Legendre
transform of g( .) (identify g(.) with F*( .) and g*(.) with F(.) in definition 1). Thus
minimizing E1 (x) with respect to x is equivalent to minimizing E1 (x, Y) = f(x) +
x . y + g* (Y) with respect to x and y. (Alternatively, we can set g* (Y) = h(Y)
in the expression for E 2(x,i/) and obtain a cost function E 2(x) = f(x) + g(x).)
Alternatively minimization over x and y gives: (i) of/ax = y to determine Xt+1 in
terms of Yt, and (ii) ag* / ay = x to determine Yt in terms of Xt which, by Property
1 of the Legendre transform is equivalent to setting y = -ag / ax. Combining these
two stages gives CCCP:
f (_)
a
ag (_)
ax Xt+1 = - ax Xt .
4
Examples of CCCP
We now illustrate CCCP by giving four examples: (i) discrete time dynamical
systems for the mean field Potts model, (ii) an EM algorithm for the elastic net,
(iii) a discrete (Sinkhorn) algorithm for solving the linear assignment problem, and
(iv) the Generalized Iterative Scaling (GIS) algorithm for parameter estimation.
Example 1.
Discrete Time Dynamical Systems for the Mean Field Potts
Model. These attempt to minimize discrete energy functions of form E[V] =
2: i ,j,a,b Tij ab Via V)b + 2: ia Bia Vi a, where the {Via} take discrete values {a, I} with
linear constraints 2:i Via = 1, Va.
Discussion. Mean field algorithms minimize a continuous effective energy E ett [S; T]
to obtain a minimum of the discrete energy E[V] in the limit as T f-7 a. The
{Sial are continuous variables in the range [0 ,1] and correspond to (approximate)
estimates of the mean states of the {Via}. As described in [12}, to ensure that the
minima of E[V] and E ett [S; T] all coincide (as T f-7 0) it is sufficient that T ijab
be negative definite. Moreover, this can be attained by adding a term -K 2: ia
to E[V] (for sufficiently large K) without altering the structure of the minima of
E[V] . Hence, without loss of generality we can consider 2: i ,j,a,b Tijab Via V)b to be a
concave function .
Vi!
We impose the linear constraints by adding a Lagrange multiplier term
2: a Pa {2: i Via - I} to the energy where the {Pa} are the Lagrange multipliers. The
effective energy becomes:
ia
i,j,a ,b
ia
a
We can then incorporate the Lagrange multiplier term into the convex part.
This gives: Evex [S] = T2: ia SialogSia + 2:aPa{2:iSia -I} and Ecave[S] =
2: i jab TijabSiaSjb + 2: ia BiaS ia ? Taking derivatives yields:
Evex [S] =
TI~~Sia + Pa
&g
&:s::~ (StH) = Bia?
&t E cave [S] = 2 2: j ,b TijabSjb + Bia? Applying eeep by setting
&:5;:e (st) gives T{l + log Sia (t + I)} + Pa = -2 2: j ,b TijabSjb(t)-
and
We solve for the Lagrange multipliers {Pal by imposing the constraints
1, Va. This gives a discrete update rule:
2:i Sia(t + 1) =
Sia (t + 1) =
e(-1/T){2 2:.J, b TijabSjb(t)+Oia}
'
.
2: c e( -1/T){2 2: j,b TijcbSjb(tl+Oi
c}
(4)
Algorithms of this type were derived in [lO}, [3} using different design principles.
Our second example relates to the ubiquitous EM algorithm. In general EM and
CCCP give different algorithms but in some cases they are identical. The EM algorithm seeks to estimate a variable f* = argmaxt log 2:{I} P(f, l), where {f}, {l} are
variables that depend on the specific problem formulation. It was shown in [4] that
this is equivalent to minimizing the following effective energy with respect to the
variables f and P(l): E ett [! , P(l)] = - ~ 2:1 P(l) log P(f, l) + ~ 2:{I} P(l) log P(l).
To apply CCCP to an effective energy like this we need either: (a) to decompose
E ett [!, P(l)] into convex and concave functions of f, P(l), or (b) to eliminate either
variable and obtain a convex concave decomposition in the remaining variable (d.
Theorem 4). We illustrate (b) for the elastic net [2]. (See Yuille and Rangarajan,
in preparation, for an illustration of (a)).
Example 2. The elastic net attempts to solve the Travelling Salesman Problem
(TSP) by finding the shortest tour through a set of cities at positions {Xi }' The
elastic net is represented by a set of nodes at positions {Ya} with variables {Sial
that determine the correspondence between the cities and the nodes of the net. Let
E el I [S, 171 be the effective energy for the elastic net, then the {y} variables can be
eliminated and the resulting Es[S] can be minimized using GGGP. (Note that the
standard elastic net only enforces the second set of linear constraints).
Discussion. The elastic net energy function can be expressed as [11]:
ia
a,b
where we impose the conditions L: a Sia = 1, V i and
i,a
L:i Sia =
1, V a.
The EM algorithm can be applied to estimate the {Ya}. Alternatively we can solve
for the {Ya} variables to obtain Yb = L:i a PabSiaXi where {Pab } = {Jab + 2')'Aab} -1.
We substitute this back into E ell [S, 171 to get a new energy Es[S] given by:
(6)
i ,j,a,b
i,a
Once again this is a sum of a concave and a convex part (the first term is concave
because of the minus sign and the fact that {Pba } and Xi . Xj are both positive semidefinite.) We can now apply GGGP and obtain the standard EM algorithm for this
problem. (See Yuille and Rangarajan, in preparation, for more details).
Our final example is a discrete iterative algorithm to solve the linear assignment
problem. This algorithm was reported by Kosowsky and Yuille in [5] where it was
also shown to correspond to the well-known Sinkhorn algorithm [9]. We now show
that both Kosowsky and Yuille's linear assignment algorithm, and hence Sinkhorn's
algorithm are examples of CCCP (after a change of variables).
Example 3. The linear assignment problem seeks to find the permutation matrix
{TI ia } which minimizes the energy E[m = L: ia TI ia A ia , where {Aia} is a set of
assignment values. As shown in [5} this is equivalent to minimizing the (convex)
Ep[P] energy given by Ep[P] = L: aPa + ~ L:i log L:a e-,B(Aia+Pa) , where the solution is given by TI;a = e-,B(Aia+Pa) / L:b e-,B(Aib+Pb) rounded off to the nearest
integer (for sufficiently large fJ). The iterative algorithm to minimize Ep[P] (which
can be re-expressed as Sinkhorn's algorithm, see [5}) is of form:
(7)
and can be re-expressed as GGGP.
Discussion. By performing the change of coordinates fJPa = - log r a V a (for r a
>
0, Va) we can re-express the Ep[P] energy as:
(8)
Observe that the first term of Er [r] is convex and the second term is concave (this
can be verified by calculating the Hessian). Applying CCCP gives the update rule:
1
rt+l =
a
2:= 2:::
i
e-,BAia
e-,BAibrt'
b
(9)
b
which corresponds to equation (7).
Example 4. The Generalized Iterative Scaling (GIS) Algorithm [ll for estimating
parameters in parallel.
Discussion. The GIS algorithm is designed to estimate the parameter
Xof a distri-
bution P(x : X) = eX.?(x) IZ[X] so that 2:::x P(x; X)?(x) = h, where h are observation data (with components indexed by j.t). It is assumed that ?fJ,(x) ::::: 0, V j.t,x,
hfJ, ::::: 0, V j.t, and 2:::fJ, ?fJ, (x) = 1, V x and 2:::fJ, hfJ, = 1. (All estimation problems of
this type can be transformed into this form [lj).
Darroch and Ratcliff [ll prove that the following GIS algorithm is guaranteed to
converge to value X* that minimizes the (convex) cost function E(X) = log Z[X]-X.h
and hence satisfies 2::: x P(x; X*)?(x) = h. The GIS algorithms is given by:
Xt+! =
Xt -
log ht
+ log h,
(10)
where ht = 2::: x P(x; Xt )?(x) {evaluate log h componentwise: (log h)fJ, = log hf),')
To show that GIS can be reformulated as CCCP, we introduce a new variable
iJ = eX (componentwise). We reformulate the problem in terms of minimizing
the cost function E,B [iJ] = log Z[log(iJ)] - h . (log iJ). A straightforward calculation shows that -h . (log iJ) is a convex function of iJ with first derivative being
-hi iJ (where the division is componentwise). The first derivative of log Z[log(iJ)] is
(II iJ) 2::: x ?(x)P(x: log ,8) (evaluated componentwise). To show that log Z[log(iJ)] is
concave requires computing its Hessian and applying the Cauchy-Schwarz inequality
using the fact that 2:::fJ, ?fJ,(x) = 1, V x and that ?fJ,(x) ::::: 0, V j.t,x. We can therefore apply CCCP to E,B [iJ] which yields l/iJH1 = l/iJt x Ilh x ht (componentwise) ,
which is GIS (by taking logs and using log ,8 = X).
5
Conclusion
CCCP is a general principle which can be used to construct discrete time iterative
dynamical systems for almost any energy minimization problem. It gives a geometric perspective on Legendre minimization (though not on Legendre min-max).
We have illustrated that several existing discrete time iterative algorithms can be reinterpreted in terms of CCCP (see Yuille and Rangarajan, in preparation, for other
examples). Therefore CCCP gives a novel way ofthinking about and classifying existing algorithms. Moreover, CCCP can also be used to construct novel algorithms.
See, for example, recent work [13] where CCCP was used to construct a double loop
algorithm to minimize the Bethe/Kikuchi free energy (which are generalizations of
the mean field free energy).
There are interesting connections between our results and those known to mathematicians. After this work was completed we found that a result similar to Theorem
2 had appeared in an unpublished technical report by D. Geman. There also are
similarities to the work of Hoang Tuy who has shown that any arbitrary closed
set is the projection of a difference of two convex sets in a space with one more
dimension. (See http://www.mai.liu.se/Opt/MPS/News/tuy.html).
Acknowledgements
We thank James Coughlan and Yair Weiss for helpful conversations. Max Welling
gave useful feedback on this manuscript. We thank the National Institute of Health
(NEI) for grant number R01-EY 12691-01.
References
[1] J.N. Darroch and D. Ratcliff. "Generalized Iterative Scaling for Log-Linear Models".
The Annals of Mathematical Statistics. Vol. 43. No.5, pp 1470-1480. 1972.
[2] R. Durbin, R. Szeliski and A.L. Yuille." An Analysis of an Elastic net Approach to
the Traveling Salesman Problem". Neural Computation. 1 , pp 348-358. 1989.
[3] LM. Elfadel "Convex potentials and their conjugates in analog mean-field optimization". Neural Computation. Volume 7. Number 5. pp. 1079-1104. 1995.
[4] R. Hathaway. "Another Interpretation of the EM Algorithm for Mixture Distributions" . Statistics and Probability Letters. Vol. 4, pp 53-56. 1986.
[5] J. Kosowsky and A.L. Yuille. "The Invisible Hand Algorithm: Solving the Assignment
Problem with Statistical Physics". Neural Networks. , Vol. 7, No.3 , pp 477-490. 1994.
[6] E. Mjolsness and C. Garrett. "Algebraic Transformations of Objective Functions".
Neural Networks. Vol. 3, pp 651-669.
[7] A. Rangarajan, S. Gold, and E. Mjolsness. "A Novel Optimizing Network Architecture with Applications" . Neural Computation, 8(5), pp 1041-1060. 1996.
[8] A. Rangarajan, A.L. Yuille, S. Gold. and E. Mjolsness." A Convergence Proof for
the Softassign Quadratic assignment Problem". In Proceedings of NIPS '96. Denver.
Colorado. 1996.
[9] R. Sinkhorn. "A Relationship Between Arbitrary Positive Matrices and Doubly
Stochastic Matrices". Ann. Math. Statist .. 35, pp 876-879. 1964.
[10] F.R. Waugh and R.M . Westervelt. "Analog neural networks with local competition:
L Dynamics and stability". Physical Review E, 47(6), pp 4524-4536. 1993.
[11] A.L. Yuille. "Generalized Deformable Models, Statistical Physics and Matching Problems," Neural Computation, 2 pp 1-24. 1990.
[12] A.L. Yuille and J.J. Kosowsky. "Statistical Physics Algorithms that Converge." Neural Computation. 6, pp 341-356. 1994.
[13] A.L. Yuille. "A Double-Loop Algorithm to Minimize the Bethe and Kikuchi Free
Energies" . Neural Computation. In press. 2002.
| 2125 |@word xof:1 seek:2 gainesville:1 decomposition:2 minus:3 liu:1 existing:5 j1:1 designed:1 update:7 smith:1 coughlan:3 hfj:2 math:1 node:2 iterates:1 cse:1 org:1 mathematical:2 direct:1 kettlewell:1 prove:3 doubly:1 tuy:2 introduce:3 becomes:2 distri:1 estimating:1 bounded:3 moreover:4 panel:5 interpreted:1 minimizes:2 mathematician:1 ag:3 transformation:2 finding:1 cave:4 ti:4 concave:20 xd:4 grant:1 positive:5 local:1 limit:1 plus:1 range:1 unique:2 enforces:1 definite:3 x3:5 procedure:7 matching:3 projection:1 ett:4 get:2 applying:4 www:1 equivalent:7 imposed:1 courtesy:1 yt:2 straightforward:1 convex:32 rule:3 aab:1 stability:1 coordinate:3 annals:1 colorado:1 designing:1 pa:6 geman:1 bottom:1 ep:4 calculate:2 news:1 mjolsness:3 decrease:3 convexity:3 miny:1 dynamic:1 depend:1 solving:2 yuille:13 division:1 basis:1 represented:1 ilh:1 univ:1 effective:6 solve:5 pab:1 statistic:2 gi:8 think:1 transform:6 tsp:1 final:1 sequence:2 eigenvalue:1 advantage:1 net:9 j2:1 loop:4 combining:1 rapidly:1 deformable:1 gold:2 competition:1 convergence:4 double:2 rangarajan:7 generating:2 converges:2 kikuchi:3 illustrate:4 ij:11 nearest:1 auxiliary:3 implemented:1 implies:2 concentrate:1 stochastic:1 generalization:1 decompose:3 opt:1 sufficiently:2 claim:1 lm:1 estimation:2 schwarz:1 city:2 minimization:10 always:1 ax:4 derived:1 potts:4 ratcliff:2 contrast:2 am:4 helpful:1 inference:1 waugh:1 el:1 eliminate:1 lj:1 transformed:1 html:1 ell:1 field:5 construct:5 once:1 eliminated:1 argmaxt:1 x4:4 identical:1 minimized:1 others:1 t2:1 report:1 national:1 cheaper:1 geometry:1 softassign:1 attempt:2 ab:1 interest:1 reinterpreted:1 mixture:1 semidefinite:1 indexed:1 iv:1 re:4 altering:1 assignment:9 cost:3 introducing:2 tour:1 pal:1 reported:1 st:1 off:1 physic:3 rounded:1 again:1 derivative:3 li:2 potential:1 summarized:1 satisfy:1 mp:1 vi:2 lot:1 closed:1 analyze:1 bution:1 hf:1 parallel:1 minimize:8 oi:1 pba:1 who:1 correspond:2 identify:1 yield:2 weak:1 emphasizes:1 email:1 definition:3 energy:28 pp:11 james:3 e2:5 proof:6 proved:1 conversation:1 ubiquitous:1 garrett:1 back:1 manuscript:1 attained:1 wei:1 yb:1 formulation:2 evaluated:1 though:1 generality:2 implicit:1 stage:1 traveling:1 hand:1 defines:2 usa:1 building:1 aib:1 multiplier:4 former:1 hence:6 illustrated:2 ll:2 generalized:6 ay:1 invisible:1 fj:10 geometrical:2 novel:3 recently:1 denver:1 jab:2 physical:1 volume:1 extend:1 analog:2 interpretation:1 imposing:1 centre:1 had:1 similarity:1 sinkhorn:5 recent:1 perspective:2 optimizing:1 phone:1 manipulation:2 inequality:1 minimum:6 impose:3 ey:1 converge:3 determine:5 shortest:1 monotonically:2 ii:3 relates:2 technical:2 af:4 calculation:2 dept:1 cccp:38 e1:7 va:3 iteration:1 preserved:1 want:1 subject:1 ufl:1 anand:3 integer:1 iii:1 xj:1 gave:1 architecture:1 inner:2 idea:1 expression:1 darroch:2 algebraic:3 reformulated:1 hessian:5 useful:1 tij:1 se:1 transforms:2 statist:1 tth:1 http:1 mai:1 sign:1 discrete:14 write:1 iz:1 vol:4 express:3 key:1 four:1 reformulation:1 pb:1 verified:1 ht:3 convert:1 sum:5 bia:3 letter:1 almost:4 scaling:5 fl:1 hi:1 guaranteed:4 distinguish:1 apa:2 correspondence:1 quadratic:1 durbin:1 constraint:9 westervelt:1 x2:6 min:4 performing:1 legendre:19 conjugate:2 describes:1 em:9 sth:1 xo:4 computationally:1 equation:1 remains:1 needed:1 travelling:1 salesman:2 ijt:1 decomposing:3 apply:4 observe:2 yair:1 aia:3 florida:1 original:1 substitute:1 top:1 remaining:1 include:1 ensure:1 completed:1 graphical:1 calculating:1 giving:2 prof:3 r01:1 objective:1 eh1:1 rt:5 gradient:2 minx:1 thank:2 street:1 mail:1 cauchy:1 sia:6 relationship:3 illustration:2 reformulate:2 minimizing:8 relate:1 negative:1 design:1 elfadel:1 ski:1 zt:2 observation:1 descent:1 arbitrary:2 nei:1 complement:2 unpublished:1 required:1 kosowsky:4 componentwise:5 connection:1 nip:1 suggested:1 proceeds:1 dynamical:6 below:2 appeared:1 max:6 oia:1 ia:12 suitable:1 eye:1 health:1 fax:2 text:1 review:1 geometric:1 acknowledgement:1 tangent:2 loss:1 permutation:1 interesting:1 hoang:1 sufficient:1 principle:2 viewpoint:1 classifying:1 lo:1 repeat:1 free:4 bias:1 allow:1 understand:2 institute:2 szeliski:1 taking:2 curve:3 dimension:1 feedback:1 concavity:2 san:1 coincide:1 welling:1 approximate:1 global:2 assumed:1 francisco:1 xi:4 alternatively:4 continuous:2 iterative:11 bethe:3 ca:1 elastic:8 tel:1 constructing:1 main:1 fillmore:1 x1:2 en:1 ff:1 tl:1 position:2 xl:2 theorem:15 xt:25 emphasized:1 specific:1 er:1 exists:1 adding:2 easier:1 saddle:3 lagrange:5 expressed:5 hathaway:1 corresponds:1 satisfies:1 goal:1 formulated:1 ann:1 room:1 cise:2 change:3 hyperplane:1 e:2 ya:3 select:1 preparation:3 incorporate:1 evaluate:1 ex:2 |
1,236 | 2,126 | Very loopy belief propagation for
unwrapping phase images
Brendan J . Freyl, Ralf Koetter2, Nemanja Petrovic 1 ,2
Probabilistic and Statistical Inference Group, University of Toronto
http://www.psi.toronto.edu
Electrical and Computer Engineering, University of Illinois at Urbana
1
2
Abstract
Since the discovery that the best error-correcting decoding algorithm can be viewed as belief propagation in a cycle-bound graph,
researchers have been trying to determine under what circumstances "loopy belief propagation" is effective for probabilistic inference. Despite several theoretical advances in our understanding of
loopy belief propagation, to our knowledge, the only problem that
has been solved using loopy belief propagation is error-correcting
decoding on Gaussian channels. We propose a new representation
for the two-dimensional phase unwrapping problem, and we show
that loopy belief propagation produces results that are superior to
existing techniques. This is an important result, since many imaging techniques, including magnetic resonance imaging and interferometric synthetic aperture radar, produce phase-wrapped images.
Interestingly, the graph that we use has a very large number of
very short cycles, supporting evidence that a large minimum cycle
length is not needed for excellent results using belief propagation.
1
Introduction
Phase unwrapping is an easily stated, fundamental problem in image processing
(Ghiglia and Pritt 1998). Each real-valued observation on a 1- or 2-dimensional
grid is measured modulus a known wavelength, which we take to be 1 without loss
of generality. Fig. Ib shows the wrapped, I-dimensional waveform obtained from the
original waveform shown in Fig. la. Every time the original waveform goes above 1
or below 0, it is wrapped to 0 or 1, respectively. The goal of phase unwrapping is to
infer the original, unwrapped curve from the wrapped measurements, using using
knowledge about which signals are more probable a priori.
In two dimensions, exact phase unwrapping is exponentially more difficult than 1dimensional phase unwrapping and has been shown to be NP-hard in general (Chen
and Zebker 2000). Fig. lc shows the wrapped output of a magnetic resonance
imaging device, courtesy of Z.-P. Liang. Notice the "fringe lines" - boundaries
across which wrappings have occurred. Fig. Id shows the wrapped terrain height
measurements from an interferometric synthetic aperture radar, courtesy of Sandia
National Laboratories, New Mexico.
(a)
(b)
(d)
Figure 1: (a) A waveform measured on a 1-dimensional grid. (b) The phase-wrapped version
ofthe waveform in (a), where the wavelength is 1. (c) A wrapped intensity ma p from a magnetic
resonance imaging device, measured on a 2-dimensional grid (courtesy of Z .-P. Liang). (d)
A wrapped topographic map measured on a 2-dimensional grid (courtesy of Sandia National
Laboratories, New Mexico) .
A sensible goal in phase unwrapping is to infer the gradient field of the original
surface. The surface can then be reconstructed by integration. Equivalently, the
goal is to infer the number of relative wrappings, or integer "shifts", between every
pair of neighboring measurements. Positive shifts correspond to an increase in the
number of wrappings in the direction of the x or y coordinate, whereas negative
shifts correspond to a decrease in the number of wrappings in the direction of the
x or y coordinate. After arbitrarily assigning an absolute number of wrappings to
one point, the absolute number of wrappings at any other point can be determined
by summing the shifts along a path connecting the two points. To account for
direction, when taking a step against the direction of the coordinate, the shift
should be subtracted.
When neighboring measurements are more likely closer together than farther apart
a priori, I-dimensional waveforms can be unwrapped optimally in time that is linear
in the waveform length. For every pair of neighboring measurements, the shift that
makes the unwrapped values as close together as possible is chosen. For example,
the shift between 0.4 and 0.5 would be 0, whereas the shift between 0.9 and 0.0
would be -1.
For 2-dimensional surfaces and images, there are many possible I-dimensional paths
between any two points. These paths should be examined in combination, since the
sum of the shifts along every such path should be equal. Viewing the shifts as state
variables, the cut-set between any two points is exponential in the size of the grid,
making exact inference for general priors NP-hard (Chen and Zebker 2000).
The two leading fully-automated techniques for phase unwrapping are the least
squares method and the branch cut technique (Ghiglia and Pritt 1998). (Some other
techniques perform better in some circumstances, but need additional information
or require hand-tweaking.) The least squares method begins by making a greedy
guess at the gradient between every pair of neighboring points. The resulting vector
field is not the gradient field of a surface, since in a valid gradient field, the sum of
the gradients around every closed loop must be zero (that is, the curl must be 0).
For example, the 2 x 2 loop of measurements 0.0, 0.3, 0.6, 0.9 will lead to gradients of
0.3,0.3,0.3, 0.1 around the loop, which do not sum to O. The least squares method
proceeds by projecting the vector field onto the linear subspace of gradient fields.
The result is integrated to produce the surface. The branch cut technique also
begins with greedy decisions for the gradients and then identifies untrustworthy
regions of the image whose gradients should not be used during integration. As
shown in our results section, both of these techniques are suboptimal.
Previously, we attempted to use a relaxed mean field technique to solve this problem
(Achan, Frey and Koetter 2001). Here, we take a new approach that works better
and is motivated by the impressive results of belief propagation in cycle-bound
graphs for error-correcting decoding (Wiberg, Loeliger and Koetter 1995; MacKay
and Neal 1995; Frey and Kschischang 1996; Kschischang and Frey 1998; McEliece,
MacKay and Cheng 1998). In contrast to other work (Ghiglia and Pritt 1998;
Chen and Zebker 2000; Koetter et al. 2001), we introduce a new framework for
quantitative evaluation, which impressively places belief propagation much closer
to the theoretical limit than other leading methods.
It is well-known that belief propagation (a.k.a. the sum-product algorithm, probability propagation) is exact in graphs that are trees (Pearl 1988), but it has been
discovered only recently that it can produce excellent results in graphs with many
cycles. Impressive results have been obtained using loopy belief propagation for
super-resolution (Freeman and Pasztor 1999) and for infering layered representations of scenes (Frey 2000) . However, despite several theoretical advances in our
understanding of loopy belief propagation (c.f. (Weiss and Freeman 2001)) and proposals for modifications to the algorithm (c.f. (Yedidia, Freeman and Weiss 2001)) ,
to our knowledge, the only problem that has been solved by loopy belief propagation
is error-correcting decoding on Gaussian channels.
We conjecture that although phase unwrapping is generally NP-hard, there exists a
near-optimal phase unwrapping algorithm for Gaussian process priors. Further, we
believe that algorithm to be loopy belief propagation.
2
Loopy Belief Propagation for Phase Unwrapping
As described above, the goal is to infer the number of relative wrappings , or integer
"shifts" , between every pair of neighboring measurements. Denote the x-direction
shift at (x,y) by a(x , y) and the y-direction shift at (x , y) by b(x , y), as shown in
Fig.2a. If the sum of the shifts around every short loop of 4 shifts (e.g., a(x,y) +
b(x + l,y) - a(x , y + 1) - b(x,y) in Fig. 2a) is zero, then perturbing a path will
not change the sum of the shifts along the path. So, a valid set of shifts S =
{a(x,y) , b(x , y) : x = 1, ... , N -1;y = 1, .. . , M -I} in an N x M image must
satisfy the constraint
a(x,y)
+ b(x + l,y)
- a(x,y
+ 1) -
b(x,y) = 0,
(1)
for x = 1, ... , N -1, Y = 1, ... , M -1. Since a(x, y) +b(x+ 1, y) -a(x, y+ 1) -b(x, y)
is a measure of curl at (x, y), we refer to (1) as a "zero-curl constraint", reflecting
the fact that the curl of a gradient field is O. In this way, phase unwrapping is
formulated as the problem of inferring the most probable set of shifts subject to
satisfying all zero-curl constraints.
We assume that given the set of shifts, the unwrapped surface is described by a loworder Gaussian process. The joint distribution over the shifts S = {a(x, y), b(x, y) :
x = 1, ... , N - 1; Y = 1, ... , M - I} and the wrapped measurements <I> = {?(x, y) :
(b)
(a)
x-direction shifts (' a's)
t.: ~-r1
X
X
X
X
'E
X [f X
X
X
X
X
t5
X
X
X
X
X
X
'6
X
X ) X
X
X
X
X
X ) X
X
X
X
X
X
X
X
X
Vi'
(x, y + l )
b(x, y )
X
a(x,y+ l )
-7
X (x+ l ,y+ l )
.<:::
I
I
<fl
<fl
c
0
b(x+ l ,y)
~
>.
(x,y)
X
-7
a(x,y)
X (x+ l ,y)
(d)
X
it2
t~
a(x, y)
it l
t
Figure 2: (a) Positive x-direction shifts (arrows labeled a) and positive y-direction shifts
(arrows labeled b) between neighboring measurements in a 2 X 2 patch of points (marked by
X 's) , (b) A graphical model that describes the zero-curl constraints (black discs) between
neighboring shift variables (white discs), 3-element probability vectors (J-L's) on the relative
shifts between neighboring variables (-1, 0, or +1) are propagated across the network: (c)
Constraint-to-shift vectors are computed from incoming shift-to-constraint vectors; (d) Shiftto-constraint vectors are computed from incoming constraint-to-shift vectors; (d) Estimates of
the marginal probabilities of the shifts given the data are computed by combining incoming
constra int-to-sh ift vectors,
0 :::; r/J(x, y)
< 1, x = 1, .. . , N; y
= 1, . . . , M } can be expressed in t he form
N- l M- l
P(S , <I? ex:
II II 5(a(x,y) +b(x +1 ,y) -
a(x,y +1 ) -b(x,y))
x=l y=l
N-l M
. II II
x= l y=l
N
e-(c/>(x+l,y)-c/>(x,y)-a(x,y))2/ 2u 2
M-l
II II
e-(C/>(x,y+1)-c/>(x,y)-b(x,y))2/ 2u 2 .
(2)
x= l y=l
The zero-curl constraints are enforced by 5 (.), which evaluates to 1 if its argument is
oand evaluates to 0 otherwise. We assume t he slope of t he surface is limited so t hat
t he unknown shifts take on t he values -1 , 0 and 1. a 2 is t he variance between two
neighboring measurements in t he unwrapped image, but we find t hat in practice it
can be estimated directly from t he wrapped image.
Phase unwrapping consists of making inferences about t he a's and b's in t he above
probability model. For example, t he marginal probability t hat t he x-direction shift
at (x,y) is k given an observed wrapped image <I> , is
P (a (x,y) = kl<I? ex:
L
P(S , <I? .
(3)
S:a(x ,y)=k
For an N x M grid, t he above sum has roughly 32N M terms and so exact inference
is intractable.
The factorization of t he joint distribution in (2) can be described by a graphical
model, as shown in Fig. 2b. In t his graph , each white disc sits on t he border between
two measurements (marked by x's), and corresponds to eit her an x-direction shift
(a's ) or a y-direction shift (b's) . Each black disc corresponds to a zero-curl constraint
(5(?) in (2)), and is connected to t he 4 shifts t hat it constrains to sum to O.
P robability propagation computes messages (3-vectors denoted by J-L) which are
passed in both directions on every edge in t he network. The elements of each 3vector correspond to t he allowed values of t he neighboring shift, -1 , 0 and 1. Each
of t hese 3-vectors can be t hought of as a probability distribut ion over t he 3 possible
values t hat t he shift can take on.
Each element in a constraint-to-shift message summarizes the evidence from the
other 3 shifts involved in the constraint, and is computed by averaging the allowed
configurations of evidence from the other 3 shifts in the constraint. For example, if
ILl ' IL2 and IL3 are 3-vectors entering a constraint as shown in Fig. 2c, the outgoing
3-vector, IL4' is computed using
1
f.t4i = L
j=-l
1
L
L
J(k
k=-ll=-l
+ l- i
- j)f.tljf.t2kf.t31,
(4)
and then normalized for numerical stability. The other 3 messages produced at the
constraint are computed in a similar fashion .
Shift-to-constraint messages are computed by weighting incoming constraint-to-shift
messages with the likelihood for the shift. For example, if ILl is a 3-vector entering
an x-direction shift as shown in Fig. 2d, the outgoing 3-vector, IL2 is computed using
(5)
and then normalized. Messages produced by y-direction shifts are computed in a
similar fashion.
At any step in the message-passing process, the messages on the edges connected
to a shift variable can be combined to produce an approximation to the marginal
probability for that shift, given the observations. For example, if ILl and IL2 are the
3-vectors entering an x-direction shift as shown in Fig. 2e, the approximation is
P(a(x , y) = il<I? = (f.tlif.t2i)/(Lf.tljf.t2j).
(6)
j
Given a wrapped image, the variance a 2 is estimated directly from the wrapped
image, the probability vectors are initialized to uniform distributions, and then
probability vectors are propagated across the graph in an iterative fashion. Different
message-passing schedules are possible, ranging from fully parallel, to a "forwardbackward-up-down" -type schedule, in which messages are passed across the network
to the right, then to the left, then up and then down. For an N x M grid, each
iteration takes O(N M) scalar computations.
After probability propagation converges (or, after a fixed number of iterations),
estimates of the marginal probabilities of the shifts given the data are computed, and
the most probable value of each shift variable is selected. The resulting configuration
of the shifts can then be integrated to obtain the unwrapped surface. If some
zero-curl constraints remain violated, a robust integration technique, such as least
squares integration (Ghiglia and Pritt 1998), should be used.
3
Experimental results
Generally, belief propagation in cycle-bound graphs is not guaranteed to converge.
Even if it does converge, the approximate marginals may not be close to the true
marginals. So, the algorithm must be verified by experiments.
On surfaces drawn from Gaussian process priors, we find that the belief propagation
algorithm produces significantly lower reconstruction errors than the least squares
method and the branch cut technique.
Here, we focus on the performances of the algorithms for real data recorded from a
synthetic aperture radar device (Fig. 1d) . Since our algorithm assumes the surface
is Gaussian given the shifts, a valid concern is that it will not perform well when
the Gaussian process prior is incorrect.
(a)
(b)
en
c
Q)
'6
CO
.....
-
0)
o
o
.....
..... 10- 1
.....
Q)
"0
~
::J
g
? 10- 2
Minimum wavelength
required for error-free
unwrapping using algs
that infer relative
~
shifts of -1 , 0 and +1
Q)
~
6
8
10
12
14
16 18 20 22 24 26
Wavelength, "-
Figure 3: (a) After 10 iterations of belief propagation using the phase-wrapped surface from
Fig. Id, hard decisions were made for the shift variables and the resulting shifts were integrated
to produce this unwrapped surface. (b) Reconstruction error versus wrapping wavelength for
our technique, the least squares method and the branch cuts technique .
Fig. 3a shows the surface that is obtained by setting (/2 to the mean squared difference between neighboring wrapped values, applying 10 iterations of belief propagation, making hard decisions for the integer shifts, and integrating the resulting
gradients. Since this is real data, we do not know the "ground truth" . However,
compared to the least squares method, our algorithm preserves more detail. The
branch cut technique is not able to unwrap the entire surface.
To obtain quantitative results on reconstruction error, we use the surface produced
by the least squares method as the "ground truth" . To determine the effect of wrapping wavelength on algorithm performance, we rewrap this surface using different
wavelengths. For each wavelength, we compute the reconstruction error for belief
propagation, least squares and branch cuts. Note that by using least squares to
obtain the ground truth, we may be biasing our results in favor of least squares.
Fig. 3b shows the logarithm of t he mean squared error in the gradient field of the
reconstructed surface as a function of the wrapping wavelength, >., on a log-scale.
(The plot for the mean squared error in t he surface heights looks similar.) As >. -+ 0,
unwrapping becomes impossible and as >. -+ 00, unwrapping becomes trivial (since
no wrappings occur), so algorithms have waterfall-shaped curves.
The belief propagation algorithm clearly obtains significantly lower reconstruction
errors. Viewed another way, belief propagation can tolerate much lower wrapping
wavelengths for a given reconstruction error. Also, it turns out that for this surface,
it is impossible for an algorithm that infers relative shifts of -1,0 and 1 to obtain a
reconstruction error of 0, unless A ::::: 12.97. Belief propagation obtains a zero-error
wavelength that is significantly closer to this limit than the least squares method
and the branch cuts technique.
4
Conclusions
Phase unwrapping is a fundamental problem in image processing and although it
has been shown to be NP-hard for general priors (Chen and Zebker 2000), we
conjecture there exists a near-optimal phase unwrapping algorithm for Gaussian
process priors. Further, we believe that algorithm to be loopy belief propagation.
Our experimental results show that loopy belief propagation obtains significantly
lower reconstruction errors compared to the least squares method and the branch
cuts technique (Ghiglia and Pritt 1998) , and performs close to the theoretical limit
for techniques that infer relative wrappings of -1, 0 and + 1. The belief propagation
algorithm runs in about the same time as the other techniques.
References
Achan, K. , Frey, B. J. , and Koetter, R. 2001. A factorized variational technique for phase
unwrapping in Markov random fields. In Uncertainty in Artificial Intelligence 2001.
Seattle, Washington.
Chen, C. W. and Zebker, H. A. 2000. Network approaches to two-dimensional phase
unwrapping: intractability and two new algorithms. Journal of the Optical Society
of America A, 17(3):401- 414.
Freeman, W. and Pasztor, E. 1999. Learning low-level vision. In Proceedings of the
Int ernational Conference on Computer Vision, pages 1182- 1189.
Frey, B. J. 2000. Filling in scenes by propagating probabilities through layers and into
appearance models. In Proceedings of th e IEEE Conference on Computer Vision and
Pattern Recognition.
Frey, B. J. and Kschischang, F. R. 1996. Probability propagation and iterative decoding. In Proceedings of the 34th Allerton Conference on Communication, Control and
Computing 1996.
Ghiglia, D. C. and Pritt, M. D. 1998. Two-Dimensional Phase Unwrapping. Theory,
Algorithms and Software. John Wiley & Sons.
Koetter, R. , Frey, B. J ., Petrovic, N., and Munson, Jr., D . C. 2001. Unwrapping phase
images by propagating probabilities across graphs. In Proceedings of the International
Conference on Acoustics, Speech and Signal Processing. IEEE Press.
Kschischang, F. R. and Frey, B. J. 1998. Iterative decoding of compound codes by probability propagation in graphical models. IEEE Journal on Selected Areas in Communications, 16(2):219- 230.
MacKay, D . J . C. and Neal, R . M. 1995. Good codes based on very sparse matrices. In
Boyd, C., editor, Cryptograph,!! and Coding. 5th IMA Conference, number 1025 in
Lecture Notes in Computer SCience, pages 100- 111. Springer, Berlin Germany.
McEliece, R. J. , MacKay, D. J . C., and Cheng, J. F. 1998. Turbo-decoding as an instance of Pearl's 'belief propagation' algorithm . IEEE Journal on Selected Areas in
Communications, 16.
Pearl , J. 1988. Probabilistic R easoning in Intelligent Systems. Morgan Kaufmann, San
Mateo CA.
Weiss, Y. and Freeman, W. 2001. On the optimaility of solutions of the max-product
belief propagation algorithm in artbitrary graphs. IEEE Transactions on Information
Theory, Special Issue on Codes on Graphs and Iterative Algorithms, 47(2):736- 744 .
Wiberg, N., Loeliger, H.-A., and Koetter, R. 1995. Codes and iterative decoding on general
graphs. European Transactions on Telecommunications, 6:513- 525.
Yedidia, J. , Freeman , W . T., and Weiss, Y. 2001. Generalized belief propagation. In
Advances in Neural Information Processing Systems 13. MIT Press, Cambridge MA.
| 2126 |@word version:1 configuration:2 loeliger:2 interestingly:1 existing:1 assigning:1 must:4 john:1 numerical:1 koetter:6 plot:1 greedy:2 selected:3 device:3 guess:1 intelligence:1 short:2 farther:1 t2j:1 loworder:1 toronto:2 sits:1 allerton:1 height:2 along:3 incorrect:1 consists:1 introduce:1 roughly:1 freeman:6 unwrapped:7 becomes:2 begin:2 factorized:1 what:1 quantitative:2 every:9 control:1 positive:3 engineering:1 frey:9 limit:3 despite:2 id:2 path:6 black:2 mateo:1 examined:1 co:1 limited:1 factorization:1 practice:1 lf:1 area:2 significantly:4 boyd:1 integrating:1 tweaking:1 onto:1 close:3 layered:1 applying:1 impossible:2 www:1 map:1 courtesy:4 go:1 resolution:1 correcting:4 his:1 ralf:1 it2:1 stability:1 coordinate:3 exact:4 element:3 satisfying:1 recognition:1 cut:9 labeled:2 observed:1 t2i:1 electrical:1 solved:2 region:1 cycle:6 connected:2 munson:1 decrease:1 forwardbackward:1 constrains:1 hese:1 radar:3 easily:1 joint:2 eit:1 america:1 effective:1 artificial:1 whose:1 valued:1 solve:1 otherwise:1 favor:1 topographic:1 propose:1 reconstruction:8 product:2 neighboring:11 loop:4 combining:1 interferometric:2 seattle:1 r1:1 produce:7 converges:1 propagating:2 measured:4 direction:16 waveform:7 viewing:1 require:1 probable:3 around:3 ground:3 wiberg:2 mit:1 clearly:1 gaussian:8 super:1 focus:1 waterfall:1 likelihood:1 contrast:1 brendan:1 inference:5 integrated:3 entire:1 her:1 germany:1 issue:1 ill:3 denoted:1 priori:2 distribut:1 resonance:3 integration:4 mackay:4 special:1 marginal:4 field:10 equal:1 shaped:1 washington:1 look:1 filling:1 np:4 intelligent:1 algs:1 preserve:1 national:2 ima:1 phase:22 message:10 evaluation:1 sh:1 edge:2 closer:3 il2:3 unless:1 tree:1 logarithm:1 initialized:1 theoretical:4 instance:1 loopy:12 unwrap:1 infering:1 uniform:1 optimally:1 synthetic:3 combined:1 petrovic:2 fundamental:2 international:1 probabilistic:3 decoding:8 connecting:1 together:2 squared:3 recorded:1 leading:2 account:1 coding:1 int:2 satisfy:1 vi:1 closed:1 parallel:1 slope:1 square:13 il:1 variance:2 kaufmann:1 correspond:3 ofthe:1 disc:4 produced:3 researcher:1 against:1 evaluates:2 involved:1 psi:1 propagated:2 knowledge:3 infers:1 schedule:2 reflecting:1 tolerate:1 wei:4 generality:1 mceliece:2 hand:1 propagation:34 believe:2 modulus:1 effect:1 normalized:2 true:1 entering:3 laboratory:2 neal:2 white:2 wrapped:16 during:1 ll:1 generalized:1 trying:1 performs:1 image:13 ranging:1 variational:1 recently:1 superior:1 perturbing:1 exponentially:1 occurred:1 he:23 marginals:2 measurement:11 refer:1 cambridge:1 curl:9 grid:7 illinois:1 impressive:2 surface:19 apart:1 compound:1 arbitrarily:1 morgan:1 minimum:2 additional:1 relaxed:1 determine:2 converge:2 signal:2 ii:6 branch:8 infer:6 circumstance:2 vision:3 iteration:4 ion:1 proposal:1 whereas:2 subject:1 integer:3 near:2 automated:1 suboptimal:1 shift:56 motivated:1 passed:2 speech:1 passing:2 generally:2 http:1 notice:1 estimated:2 group:1 drawn:1 verified:1 achan:2 imaging:4 graph:12 sum:8 enforced:1 run:1 uncertainty:1 telecommunication:1 place:1 patch:1 decision:3 summarizes:1 bound:3 fl:2 layer:1 guaranteed:1 easoning:1 cheng:2 turbo:1 occur:1 constraint:18 scene:2 software:1 argument:1 optical:1 conjecture:2 freyl:1 combination:1 jr:1 across:5 describes:1 remain:1 son:1 making:4 modification:1 projecting:1 previously:1 turn:1 needed:1 know:1 sandia:2 yedidia:2 magnetic:3 subtracted:1 hat:5 original:4 assumes:1 graphical:3 society:1 wrapping:13 gradient:12 subspace:1 berlin:1 sensible:1 trivial:1 length:2 code:4 mexico:2 liang:2 difficult:1 equivalently:1 stated:1 negative:1 unknown:1 perform:2 observation:2 markov:1 urbana:1 pasztor:2 supporting:1 communication:3 discovered:1 intensity:1 pair:4 required:1 kl:1 acoustic:1 pearl:3 able:1 proceeds:1 below:1 pattern:1 biasing:1 including:1 max:1 belief:29 nemanja:1 identifies:1 prior:6 understanding:2 discovery:1 relative:6 loss:1 fully:2 lecture:1 impressively:1 versus:1 editor:1 intractability:1 unwrapping:22 ift:1 free:1 taking:1 absolute:2 sparse:1 curve:2 dimension:1 boundary:1 valid:3 t5:1 computes:1 made:1 san:1 transaction:2 reconstructed:2 approximate:1 obtains:3 aperture:3 incoming:4 summing:1 terrain:1 iterative:5 channel:2 robust:1 ca:1 untrustworthy:1 kschischang:4 excellent:2 european:1 arrow:2 border:1 allowed:2 constra:1 fig:14 en:1 fashion:3 wiley:1 lc:1 inferring:1 exponential:1 ib:1 weighting:1 down:2 oand:1 evidence:3 concern:1 exists:2 intractable:1 chen:5 wavelength:11 likely:1 appearance:1 expressed:1 scalar:1 springer:1 corresponds:2 truth:3 ma:2 fringe:1 viewed:2 goal:4 formulated:1 marked:2 hard:6 change:1 determined:1 averaging:1 experimental:2 la:1 attempted:1 il4:1 violated:1 outgoing:2 ex:2 |
1,237 | 2,127 | On the Concentration of Spectral
Properties
John Shawe-Taylor
Royal Holloway, University of London
N ella Cristianini
BIOwulf Technologies
[email protected]
nello@support-vector. net
Jaz Kandola
Royal Holloway, University of London
[email protected]
Abstract
We consider the problem of measuring the eigenvalues of a randomly drawn sample of points. We show that these values can be
reliably estimated as can the sum of the tail of eigenvalues. Furthermore, the residuals when data is projected into a subspace is
shown to be reliably estimated on a random sample. Experiments
are presented that confirm the theoretical results.
1
Introduction
A number of learning algorithms rely on estimating spectral data on a sample of
training points and using this data as input to further analyses. For example in
Principal Component Analysis (PCA) the subspace spanned by the first k eigenvectors is used to give a k dimensional model of the data with minimal residual,
hence forming a low dimensional representation of the data for analysis or clustering. Recently the approach has been applied in kernel defined feature spaces
in what has become known as kernel-PCA [5]. This representation has also been
related to an Information Retrieval algorithm known as latent semantic indexing,
again with kernel defined feature spaces [2].
Furthermore eigenvectors have been used in the HITS [3] and Google's PageRank [1]
algorithms. In both cases the entries in the eigenvector corresponding to the maximal eigenvalue are interpreted as authority weightings for individual articles or web
pages.
The use of these techniques raises the question of how reliably these quantities can
be estimated from a random sample of data, or phrased differently, how much data is
required to obtain an accurate empirical estimate with high confidence. Ng et al. [6]
have undertaken a study of the sensitivity of the estimate of the first eigenvector to
perturbations of the connection matrix. They have also highlighted the potential
instability that can arise when two eigenvalues are very close in value, so that their
eigenspaces become very difficult to distinguish empirically.
The aim of this paper is to study the error in estimation that can arise from the
random sampling rather than from perturbations of the connectivity. We address
this question using concentration inequalities. We will show that eigenvalues estimated from a sample of size m are indeed concentrated, and furthermore the sum
of the last m - k eigenvalues is subject to a similar concentration effect, both results of independent mathematical interest. The sum of the last m - k eigenvalues
is related to the error in forming a k dimensional PCA approximation, and hence
will be shown to justify using empirical projection subspaces in such algorithms as
kernel-PCA and latent semantic kernels.
The paper is organised as follows. In section 2 we give the background results and
develop the basic techniques that are required to derive the main results in section
3. We provide experimental verification of the theoretical findings in section 4,
before drawing our conclusions.
2
Background and Techniques
We will make use of the following results due to McDiarmid. Note that lEs is the
expectation operator under the selection of the sample.
TheoreIll 1 (McDiarmid!4}) Let Xl, ... ,Xn be independent random variables taking values in a set A, and assume that f : An -+~, and fi : An- l -+ ~ satisfy for
l:::;i:::;n
Xl,??? , Xn
TheoreIll 2 (McDiarmid!4}) Let Xl, ... ,Xn be independent random variables taking values in a set A, and assume that f : An -+ ~, for 1 :::; i :::; n
sup
If(xI, ... , xn) - f(XI, ... , Xi -
I,
Xi, Xi+!,???, xn)1 :::; Ci,
We will also make use of the following theorem characterising the eigenvectors of a
symmetric matrix.
TheoreIll 3 (Courant-Fischer MiniIllax TheoreIll) If M E
ric, then for k = 1, ... , m,
Ak(M) =
max
v'Mv
min - - =
vlv
dim(T) = k O#v ET
min
max
dim(T) = m - k+IO#v E T
~mxm
is symmet-
v'Mv
vlv '
with the extrama achieved by the corresponding eigenvector.
The approach adopted in the proofs of the next section is to view the eigenvalues as
the sums of squares of residuals. This is applicable when the matrix is positive semidefinite and hence can be written as an inner product matrix M = XI X, where XI is
the transpose of the matrix X containing the m vectors Xl, . . . , Xm as columns. This
is the finite dimensional version of Mercer's theorem, and follows immediately if we
take X = V VA, where M = VA VI is the eigenvalue decomposition of M. There
may be more succinct ways of representing X, but we will assume for simplicity (but
without loss of generality) that X is a square matrix with the same dimensions as
M. To set the scene, we now present a short description of the residuals viewpoint.
The starting point is the singular value decomposition of X = U~V', where U and
V are orthonormal matrices and ~ is a diagonal matrix containing the singular
values (in descending order). We can now reconstruct the eigenvalue decomposition
of M = X' X = V~U'U~V' = V AV', where A = ~2. But equally we can construct
a matrix N = XX' = U~V'V~U' = UAU' , with the same eigenvalues as M.
As a simple example consider now the first eigenvalue, which by Theorem 3 and the
above observations is given by
v'Nv
max - O,t:vEIR = v'v
A1(M)
= max
O,t:vEIR=
m
max
O,t:vEIR =
v'XX'v
v'v
max
O,t:vE IR =
m
L
IIPv(xj)11 2 =
j=l
v'v
m
L
IIxjl12 -
j=l
min
O,t:vEIR=
L
IIP;-(xj)11 2
j=l
where Pv(x) (Pv..l (x)) is the projection of x onto the space spanned by v (space
perpendicular to v), since IIxI1 2 = IIPv(x)11 2+ IIPv
..l(x)112. It follows that the first
eigenvector is characterised as the direction for which sum of the squares of the
residuals is minimal.
Applying the same line of reasoning to the first equality of Theorem 3, delivers the
following equality
m
Ak =
max
L
min
dim(V) = k O,t:vEV .
IlPv(xj)112.
(1)
J=l
Notice that this characterisation implies that if v k is the k-th eigenvector of N, then
m
L
(2)
IlPv k (xj)112,
j=l
which in turn implies that if Vk is the space spanned by the first k eigenvectors,
then
Ak =
k
L
m
Ai =
L
m
IIPVk (Xj) 112 =
j=l
i=l
L
j=l
m
IIXj W-
L
IIP'* (Xj) 11 2,
(3)
j=l
where Pv(x) (PV(x)) is the projection of x into the space V (space perpendicular
to V). It readily follows by induction over the dimension of V that we can equally
characterise the sum of the first k and last m - k eigenvalues by
m
max
i= l
m
L
dim(V) = k .
m
IIPv(xj)11 2
=
J=l
m
L
L
.
m
IIxjl12 -
)= 1
k
L
min
L
dim(V) = k .
IIPv(xj)11 2,
)= 1
m
L
(4)
IIXjl12 - . Ai = dim(V)=k
min
IlPv(xj)112.
.
.
J=l
.=1
J=l
Hence, as for the case when k = 1, the subspace spanned by the first k eigenvalues
is characterised as that for which the sum of the squares of the residuals is minimal.
Frequently, we consider all of the above as occurring in a kernel defined feature
space, so that wherever we have written Xj we should have put ?>(Xj), where ?> is
the corresponding projection.
3
Concentration of eigenvalues
The previous section outlined the relatively well-known perspective that we now
apply to obtain the concentration results for the eigenvalues of positive semi-definite
matrices. The key to the results is the characterisation in terms of the sums of
residuals given in equations (1) and (4).
Theorem 4 Let K(x,z) be a positive semi-definite kernel function on a space X,
and let J-t be a distribution on X. Fix natural numbers m and 1 :::; k < m and let
S = (Xl"'" x m) E xm be a sample of m points drawn according to J-t. Th en for
all f > 0,
P{I~ )..k(S)
-lEs[~ )..k(S)ll 2: f} :::;
2exp (
-~:m)
,
where )..k (S) is the k-th eigenvalue of the matrix K(S) with entries K(S)ij
K(Xi,Xj) and R2 = maxx Ex K(x,x).
Proof: The result follows from an application of Theorem 1 provided
1
1
2
sup 1- )..k(S) - - )..k(S \ {xd)1 :::; Rim.
s m
m
Let S = S \ {Xi} and let V (11) be the k dimensional subspace spanned by the first
k eigenvectors of K(S) (K(S)). Using equation (1) we have
m
m
D
Surprisingly a very similar result holds when we consider the sum of the last m - k
eigenvalues.
Theorem 5 Let K(x, z) be a positive semi-definite kernel function on a space X,
and let J-t be a distribution on X. Fix natural numbers m and 1 :::; k < m and let
S = (Xl, ... , Xm) E xm be a sample of m points drawn according to J-t. Then for
all f > 0,
P{I~ )..>k(S)
-lEs [~ )..>k(S)ll 2: f} :::;
2 exp (
-~:m)
,
where )..>k(S) is the sum of all but the largest k eigenvalues of the matrix K(S) with
entries K(S)ij = K(Xi,Xj) and R2 = maxxEX K(x,x).
Proof: The result follows from an application of Theorem 1 provided
sup
1~)..>k(S)
s m
-
~)..>k(S \ {xd)1 :::; R2/m.
m
Let S = S \ {xd and let V (11) be the k dimensional subspace spanned by the first
k eigenvectors of K(S) (K(S)). Using equation (4) we have
m
j=l
#i
m
#i
D
j=l
Our next result concerns the concentration of the residuals with respect to a fixed
subspace. For a subspace V and training set S , we introduce the notation
1 m
Fv(S) = -
L
IIPV(Xi )112 .
m i=l
TheoreIll 6 Let J-t be a distribution on X. Fix natural numbers m and a subspace
V and let S = (Xl, .. . , Xm) E xm be a sample of m points drawn according to J-t.
Then for all t > 0,
P{IFv(S) -lEs [Fv(S)ll
~ t} ::::: 2exp (~~r;) .
Proof: The result follows from an application of Theorem 2 provided
sup IFv(S) - F(S \ {xd U {xi)1 ::::: R2/m.
S,X i
Clearly the largest change will occur if one of the points Xi and Xi is lies in the
subspace V and the other does not. In this case the change will be at most R2/m.
D
4
Experiments
In order to test the concentration results we p erformed experiments with the Breast
cancer data using a cubic polynomial kernel. The kernel was chosen to ensure that
the spectrum did not decay too fast.
We randomly selected 50% of the data as a 'training' set and kept the remaining
50% as a 'test' set. We centered the whole data set so that the origin of the feature
space is placed at the centre of gravity of the training set. We then performed an
eigenvalue decomposition of the training set. The sum of the eigenvalues greater
than the k-th gives the sum of the residual squared norms of the training points
when we project onto the space spanned by the first k eigenvectors. Dividing this by
the average of all the eigenvalues (which measures the average square norm of the
training points in the transformed space) gives a fraction residual not captured in
the k dimensional projection. This quantity was averaged over 5 random splits and
plotted against dimension in Figure 1 as the continuous line. The error bars give
one standard deviation. The Figure la shows the full spectrum, while Figure 1b
shows a zoomed in subwindow. The very tight error bars show clearly the very tight
concentration of the sums of tail of eigenvalues as predicted by Theorem 5.
In order to test the concentration results for subsets we measured the residuals of
the test points when they are projected into the subspace spanned by the first k
eigenvectors generated above for the training set. The dashed lines in Figure 1 show
the ratio of the average squares of these residuals to the average squared norm of the
test points. We see the two curves tracking each other very closely, indicating that
the subspace identified as optimal for the training set is indeed capturing almost
the same amount of information in the test points.
5
Conclusions
The paper has shown that the eigenvalues of a positive semi-definite matrix generated from a random sample is concentrated. Furthermore the sum of the last m - k
eigenvalues is similarly concentrated as is the residual when the data is projected
into a fixed subspace.
0.7,------,-------,-------,------,-------,-------,------,,------,
0.6
0.5
0.2
0.1
Projection Dimensionality
(a)
0.14,-----,-----,-----,-----,-----,-----,-----,-----,-----,-----,
0.12
0.1
e0.08
W
1
Cii
OJ
:g
en
\
&! 0.06
0.04
0.02
'1- -
I
- -:E-- -I- --:1:- _ '.[ __
O~--L---~--~--~---L-=~~~~======~~
o
10
20
30
40
50
60
Projection Dimensionality
70
80
90
100
(b)
Figure 1: Plots ofthe fraction of the average squared norm captured in the subspace
spanned by the first k eigenvectors for different values of k. Continuous line is
fraction for training set, while the dashed line is for the test set. (a) shows the full
spectrum, while (b) zooms in on an interesting portion.
Experiments are presented that confirm the theoretical predictions on a real world
dataset. The results provide a basis for performing PCA or kernel-PCA from a
randomly generated sample, as they confirm that the subset identified by the sample
will indeed 'generalise' in the sense that it will capture most of the information in
a test sample.
Further research should look at the question of how the space identified by a subsample relates to the eigenspace of the underlying kernel operator.
References
[1] S. Brin and L. Page. The anatomy of a large-scale hypertextual (web) search engine. In Proceedings of the Seventh International World Wide Web Conference,
1998.
[2] Nello Cristianini, Huma Lodhi, and John Shawe-Taylor. Latent semantic kernels
for feature selection. Technical Report NC-TR-00-080, NeuroCOLT Working
Group, http://www.neurocolt.org, 2000.
[3] J. Kleinberg. Authoritative sources in a hyperlinked environment. In Proceedings
of 9th ACM-SIAM Symposium on Discrete Algorithms, 1998.
[4] C. McDiarmid. On the method of bounded differences. In Surveys in Combinatorics 1989, pages 148- 188. Cambridge University Press , 1989.
[5] S. Mika, B. SchCilkopf, A. Smola, K.-R. MUller, M. Scholz, and G. Ratsch.
Kernel PCA and de-noising in feature spaces. In Advances in Neural Information
Processing Systems 11, 1998.
[6] Andrew Y. Ng, Alice X. Zheng, and Michael 1. Jordan. Link analysis, eigenvectors and stability. In To appear in the Seventeenth International Joint Conference on Artificial Intelligence (UCAI-Ol), 2001.
| 2127 |@word version:1 polynomial:1 norm:4 lodhi:1 decomposition:4 tr:1 jaz:2 written:2 readily:1 john:3 plot:1 intelligence:1 selected:1 short:1 authority:1 mcdiarmid:4 org:1 mathematical:1 become:2 symposium:1 introduce:1 indeed:3 frequently:1 ol:1 provided:3 estimating:1 xx:2 notation:1 project:1 underlying:1 eigenspace:1 bounded:1 what:1 interpreted:1 eigenvector:5 finding:1 iipvk:1 xd:4 gravity:1 hit:1 uk:2 appear:1 before:1 positive:5 io:1 ak:3 mika:1 alice:1 scholz:1 perpendicular:2 averaged:1 seventeenth:1 definite:4 empirical:2 maxx:1 projection:7 confidence:1 onto:2 close:1 selection:2 operator:2 noising:1 put:1 applying:1 instability:1 descending:1 www:1 starting:1 survey:1 simplicity:1 immediately:1 spanned:9 orthonormal:1 stability:1 origin:1 capture:1 environment:1 cristianini:2 raise:1 tight:2 basis:1 joint:1 differently:1 fast:1 london:2 artificial:1 vlv:2 drawing:1 reconstruct:1 fischer:1 highlighted:1 ifv:2 eigenvalue:25 hyperlinked:1 net:1 maximal:1 product:1 zoomed:1 description:1 derive:1 develop:1 ac:2 andrew:1 measured:1 ij:2 dividing:1 c:2 predicted:1 implies:2 direction:1 anatomy:1 closely:1 centered:1 brin:1 fix:3 hold:1 exp:3 estimation:1 applicable:1 maxxex:1 largest:2 clearly:2 aim:1 rather:1 vk:1 sense:1 dim:6 transformed:1 construct:1 ng:2 sampling:1 look:1 report:1 randomly:3 kandola:1 ve:1 zoom:1 individual:1 interest:1 zheng:1 semidefinite:1 accurate:1 eigenspaces:1 taylor:2 plotted:1 e0:1 theoretical:3 minimal:3 column:1 measuring:1 deviation:1 entry:3 subset:2 seventh:1 too:1 international:2 sensitivity:1 siam:1 michael:1 connectivity:1 again:1 squared:3 containing:2 uau:1 iip:2 potential:1 de:1 satisfy:1 combinatorics:1 mv:2 vi:1 performed:1 view:1 sup:4 portion:1 square:6 ir:1 iixjl12:3 ofthe:1 against:1 proof:4 dataset:1 dimensionality:2 rim:1 courant:1 generality:1 furthermore:4 smola:1 working:1 web:3 google:1 effect:1 hence:4 equality:2 symmetric:1 semantic:3 ll:3 delivers:1 characterising:1 theoreill:5 reasoning:1 recently:1 fi:1 empirically:1 tail:2 cambridge:1 ai:2 outlined:1 similarly:1 centre:1 shawe:2 perspective:1 inequality:1 muller:1 captured:2 greater:1 cii:1 dashed:2 semi:4 relates:1 full:2 technical:1 retrieval:1 equally:2 a1:1 va:2 prediction:1 basic:1 breast:1 expectation:1 mxm:1 kernel:14 achieved:1 background:2 ratsch:1 singular:2 source:1 nv:1 subject:1 jordan:1 split:1 xj:13 identified:3 inner:1 pca:7 iixj:1 eigenvectors:10 characterise:1 amount:1 concentrated:3 http:1 notice:1 estimated:4 discrete:1 group:1 key:1 drawn:4 characterisation:2 undertaken:1 kept:1 fraction:3 sum:14 almost:1 ric:1 iipv:6 capturing:1 distinguish:1 hypertextual:1 occur:1 scene:1 phrased:1 kleinberg:1 min:6 performing:1 relatively:1 according:3 wherever:1 indexing:1 equation:3 turn:1 adopted:1 apply:1 spectral:2 clustering:1 ensure:1 remaining:1 question:3 quantity:2 concentration:9 biowulf:1 diagonal:1 subspace:14 link:1 neurocolt:2 nello:2 induction:1 ratio:1 nc:1 difficult:1 reliably:3 av:1 observation:1 finite:1 perturbation:2 required:2 connection:1 engine:1 fv:2 huma:1 address:1 bar:2 xm:6 pagerank:1 royal:2 max:8 oj:1 natural:3 rely:1 residual:13 representing:1 technology:1 ella:1 loss:1 interesting:1 organised:1 authoritative:1 verification:1 article:1 mercer:1 viewpoint:1 cancer:1 surprisingly:1 last:5 transpose:1 placed:1 generalise:1 wide:1 taking:2 curve:1 dimension:3 xn:5 world:2 subwindow:1 projected:3 iixi1:1 confirm:3 xi:14 symmet:1 spectrum:3 continuous:2 latent:3 search:1 did:1 main:1 ilpv:3 whole:1 subsample:1 arise:2 succinct:1 en:2 cubic:1 pv:4 xl:7 lie:1 weighting:1 theorem:10 rhul:2 r2:5 decay:1 concern:1 ci:1 occurring:1 forming:2 vev:1 tracking:1 acm:1 change:2 characterised:2 justify:1 principal:1 experimental:1 la:1 indicating:1 holloway:2 support:1 ex:1 |
1,238 | 2,128 | Online Learning with Kernels
Jyrki Kivinen
Alex J. Smola
Robert C. Williamson
Research School of Information Sciences and Engineering
Australian National University
Canberra, ACT 0200
Abstract
We consider online learning in a Reproducing Kernel Hilbert Space. Our
method is computationally efficient and leads to simple algorithms. In
particular we derive update equations for classification, regression, and
novelty detection. The inclusion of the -trick allows us to give a robust
parameterization. Moreover, unlike in batch learning where the -trick
only applies to the -insensitive loss function we are able to derive general trimmed-mean types of estimators such as for Huber?s robust loss.
1 Introduction
While kernel methods have proven to be successful in many batch settings (Support Vector
Machines, Gaussian Processes, Regularization Networks) the extension to online methods
has proven to provide some unsolved challenges. Firstly, the standard online settings for
linear methods are in danger of overfitting, when applied to an estimator using a feature
space method. This calls for regularization (or prior probabilities in function space if the
Gaussian Process view is taken).
Secondly, the functional representation of the estimator becomes more complex as the number of observations increases. More specifically, the Representer Theorem [10] implies
that the number of kernel functions can grow up to linearly with the number of observations. Depending on the loss function used [15], this will happen in practice in most cases.
Thereby the complexity of the estimator used in prediction increases linearly over time (in
some restricted situations this can be reduced to logarithmic cost [8]).
Finally, training time of batch and/or incremental update algorithms typically increases superlinearly with the number of observations. Incremental update algorithms [2] attempt to
overcome this problem but cannot guarantee a bound on the number of operations required
per iteration. Projection methods [3] on the other hand, will ensure a limited number of
updates per iteration. However they can be computationally expensive since they require
one matrix multiplication at each step. The size of the matrix is given by the number of
kernel functions required at each step.
Recently several algorithms have been proposed [5, 8, 6, 12] performing perceptron-like
updates for classification at each step. Some algorithms work only in the noise free case,
others not for moving targets, and yet again others assume an upper bound on the complexity of the estimators. In the present paper we present a simple method which will allows
the use of kernel estimators for classification, regression, and novelty detection and which
copes with a large number of kernel functions efficiently.
2 Stochastic Gradient Descent in Feature Space
"! #
%$&
&$
to be studied in
Reproducing Kernel Hilbert Space The class of functions
this paper are elements of an RKHS . This means that there exists a kernel
and a dot product
such that 1)
(reproducing property); 2) is
the closure of the span of all
with
. In other words, all
are linear
combinations of kernel functions.
' '()!*
Typically
is used as a regularization functional. It is the ?length of the weight
vector in feature space? as commonly used in SV algorithms. To state our algorithm we
need to compute derivatives of functionals defined on .
/.
0
/.
!21 ' '( we obtain 3546+-, ! . More general versions of
For the regularizer +-,
+-, /. !87 ' ' lead to 3 4 +-, /.( !97: ' ';' '=< 1 .
/.B ! we compute the derivative by using the reproFor the evaluation functional >@?A,
ducing property
of G
and
obtain
3 4 > ? , /. !%3 4
)!C
# . Consequently for
E
F
H
G
I
a function
which is differentiable in its third argument we obtain
D
3 4 D # KJL # !MD: # JL K
# . Below D will be the loss function.
Regularized Risk Functionals and
standard learning setting we are
#LLearning
N J N $OPQInG the
drawn
according to some underlying
supplied with pairs
of
observations
# J . Our aim is to predict the likely outcome
distribution R
at location . Several
# J may change over time, (ii) theJ training
# N KJ N
variants are possible: (i) R
sample
may be the next observation on which to predict which leads to a true online setting, or
(iii) we may want to find an algorithm which approximately minimizes a regularized risk
functional on a given training set.
D
SG%TGU
K J V;VV #/W KJ W J R # J
1 1
W
XZY\[^] , /. ! ` _ a D #cN KJ N #cN K
(1)
Nb
1
or, in order to avoid overly complex
hypotheses,
minimize the empirical risk plus an addi/.
tional regularization term +-, . This sum is known as the regularized risk
W
_
a
X"d#Y\e , /.f ! X"Y\[^] , /.5gih +-, /. ! ` D /N J N /N gkh +-, /. for hQl9m V (2)
Njb
1
We assume that we want to minimize a loss function
which penalizes
the deviation between an observation at location and the prediction
, based on
. Since
is unknown, a standard approach is to
observations
instead minimize the empirical risk
Common loss functions are the soft margin loss function [1] or the logistic loss for classification and novelty detection [14], the quadratic loss, absolute loss, Huber?s robust loss [9],
or the -insensitive loss [16] for regression. We discuss these in Section 3.
n
In some cases the loss function depends on an additional parameter such as the width of the
margin or the size of the -insensitive zone. One may make these variables themselves
parameters of the optimization problem [15] in order to make the loss function adaptive to
the amount or type of noise present in the data. This typically results in a term
or
added to
.
D # KJL #
X #d Y\e , /.
o n
Stochastic Approximation In order to find a good estimator we would like to minimize
. This can be costly if the number of observations is large. Recently several gradient
descent algorithms for minimizing such functionals efficiently have been proposed [13, 7].
Below we extend these methods to stochastic gradient descent by approximating
Xpd#Y\e , /.
X ,
f. ! D # K J #
gkh +-, /.
(3)
.
X
and then performing
with respect to
, . Here is
either randomly
_ ;VVV `gradient
or it isdescent
chosen from X
the
new
training
instance
observed
at time . Consequently
,
. with respect to is
the gradient of
3 4 X ,
. ! D : # J # K
; gSh 3 4 +-, /. ! D : J # K
; gSh/ V (4)
/. ! 1 ' ' ( . Analogous results hold for general +-, /. !
The
equality holds if +-,
7 ' last
' . The the update equations are( hence straightforward:
S
o 354 X ,
. V
(5)
$
Here
is the learning rate controlling h the size of updates undertaken at each iteration. We will return to the issue of adjusting at a later stage.
/.
Descent Algorithm For simplicity, assume that +-, !H1 ' '( . In this case (5) becomes
o D : KJ #
K
#
gihc E! _ o h o ( D : J #
K
; V (6)
by
While (6) is convenient to use for a theoretical analysis, it is not directly amenable to
computation. For this purpose we have to express as a kernel expansion
E! a N
# cN
N
where the
N are (previously seen) training patterns. Then (6) becomes
_ o h
J # K
:
o
D
m
#
for !
N ! o _ o D : h KJ N
for ! V
(7)
(8)
(9)
(10)
Eq. (8) means that at each iteration the kernel expansion may grow by one term. Furthermore, the cost
for training
at each step is not larger than the prediction cost:
have
once
we
computed
, is obtained by the value of the derivative of at
.
#
D # J # K
N
_ _ o
Instead
all coefficients we may simply cache the power series
h _ o ofh updating
h
_
( o "!@V;VV and pick suitable terms as needed. This is particularly
_ m useful
_
if the derivatives of the loss function D will only assume discrete values, say o as
is the case when using the soft-margin type loss functions (see Section 3).
Truncation The problem with (8) and (10) is that without any further measures, the number of basis functions # will grow without bound. This is not desirable since # determines
the amount of computation needed for prediction.
The regularization term helps us here. At
are shrunk
each iteration the coefficients with
by
. Thus after $ iterations
the coefficient will be reduced to
"% . Hence:
_ o h
# K with its first derivaProposition 1 (Truncation Error) For a loss function D JL
tive bounded by & and a kernel
with
norm '
; ;(' '*) , the truncation error
N frombounded
in incurred by
dropping
terms
the
kernel
expansion
of after $ update steps is
_ o h % &) . Furthermore,
bounded by
the total truncation error by dropping all terms
which are at least $ steps old is bounded by
a <.% _ h < N
d
+
,
(11)
' o
' ' Nb o &)0/ h < 1 _ o h % &)
1
N
N
_ o! h N
N
/N
< 1
h
requirements for the
The regularization parameter can thus be used to control the
storage
R
K
5
J
expansion. In addition, it naturally allows for distributions
that
change over time in
# N J N that are much older
which cases it is desirable to forget instances
than the average
d+ ,
!
Nb
Here
. Obviously the approximation quality increases expo.% of terms retained.
nentially with the number
time scale of the distribution change [11].
3 Applications
# ! g
We now proceed to applications of (8) and (10) to specific learning situations. We utilize
the standard addition of the constant offset to the function expansion,
i.e.
.
where
and
. Hence we also update into
O$
D JL !
$
,.
o 3 X
m _
o J
N
KK __
o
hh
o
Classification A typical loss function in SVMs is the soft margin, given by
. In this situation the update equations become
n
N N g N
N Jm Z J
J #
_
/
if
otherwise.
(12)
n
In classification with the -trick we avoid having to fix the margin by treating it as a
variable [15]. The value of is found automatically by using the loss function
D # JL # !
m n oJ
K o n
(13)
m ' _ is another parameter. Since has a much clearer intuitive meaning than
where '
h
n , it is easier to tune. On the other hand, one can show [15] that the specific choice
of
h
_
has no influence on the estimate in -SV classification. Therefore we may set !
and
obtain
N = n5 __ o NN KJm N " g J N n g _ o if J
/kn (14)
o = np
o
otherwise.
# #
Finally, if we choose the hinge-loss, D JL
!
m ;o
J
N KK __
o
hh NN Jm N Z g J N if J # / m
(15)
o
otherwise.
h ! m recovers the kernel-perceptron algorithm. For nonzero h we obtain the
Setting
kernel-perceptron with regularization.
Novelty Detection The results for novelty detection [14] are similar in spirit. The setting is most useful here particularly where the estimator acts as a warning device (e.g.
network intrusion detection) and we would like to specify an upper limit on the frequency
of alerts
(/ . The relevant loss function is
where
and usually [14] one uses
rather than
in order to avoid trivial
solutions. The update equations are
# Un
#
m n o # K o n
8$
Tg D KJc Z$i!
N n5 __ o NN Zm Kn g _ o K if /8n
(16)
o Kn o
otherwise.
Considering the update of n we can see that on average only a fraction of observations
fN
will be considered for updates. Thus we only have to store a small fraction of the .
Regression We consider the following four settings: squared loss, the -insensitive loss
using the -trick, Huber?s robust loss function, and trimmed mean estimators. For con where
venience we will only use estimates
rather than
. The
extension to the latter case is straightforward. We begin with squared loss where is given
by
the update equation is
Consequently
(17)
This means that we have to store every observation we make, or more precisely, the
prediction error we made on the observation. The -insensitive loss
avoids this problem but introduces a new parameter in turn ?
the width of the insensitivity zone . By making a variable of the optimization problem
we have
The update equations now have to be
stated in terms of , and which is allowed to change during the optimization process.
This leads to
if
otherwise.
(18)
This means that every time the prediction error exceeds , we increase the insensitivity
zone by . Likewise, if it is smaller than , the insensitive zone is decreased by
.
Next let us analyze the case of regression with Huber?s robust loss. The loss is given by
*$
8!
D # JL P
! (1 J o N K ( V _ h N #
o J o
K V
m QJ o # /o @
m J o #
D JL # N K!
N __ o h h NN m J
o "o
g
$%
D
D KJc # !
\o g 5V
o g _ o
J o l
_ o
J o #
#@o (1 if J o #
(19)
1( J o K ( otherwise.
# .
As before we obtain update equations by computing the derivative of D with respect to
N __ o NN J o # K if J o # l
(20)
o < 1 J o
K otherwise.
D JL # K!
Comparing (20) with (18) leads to the question whether might not also be adjusted
adaptively. This is a desirable goal since we may not know the amount of noise present in
the data. While the -setting allowed us to form such adaptive estimators for batch learning
with the -insensitive loss, this goal has proven elusive for other estimators in the standard
batch setting. In the online situation, however, such an extension is quite natural (see also
[4]). All we need to do is make a variable of the optimization problem and set
N
K _
o
K _ o
N g _
N < 1 J J o o # o o
4 Theoretical Analysis
m n o8J # K
Wb X
.
,
1
J o l
if
otherwise.
(21)
D KJc # !
Consider now the classification problem with the soft margin
loss
denote
; here is a fixed margin
parameter.
Let
the
hypothesis of
the online algorithm
after
seeing
the
first
observations.
Thus,
at
time
, the algorithm
.
, receives the correct outcome
, and upreceives an input , makes
its
prediction
dates
into according to (5). We now wish to bound the cumulative risk
its hypothesis
. The
motivation for such bounds is roughly
follows. Assume there is
areasdrawn,
some fixed distribution from which the examples
and define
4 X , /.
n
o _
1
# J
RX /.f
, ! ?
! , D # JL K .5gk h +-, /. V
J
` X
"
`
+ E ! (1 ' '(
!
,
.Kg
Then it would be desirable for the online hypothesis
to converge towards
"
arg min
. If we can show that the cumulative risk is asymptotically
"
#
, we see that at least in some sense does converge to .
Hence, as a first step in our convergence analysis, we obtain an upper bound for the cumulative risk. In all the bounds of this section we assume
.
K# K J W b
lCm
1
' '
W
W
a X
. a X
, ' b
b
1
1
#
*! ` 1 (;
, L
. g ) ` 1 ( g _ V
(
Theorem 1 Let
be an example sequence such that
' ) for all
. Fix
)
, and choose the learning rate
. Then for any such that
'
we have
(22)
Notice that the bound does not depend on any probabilistic assumptions. If the example
sequence is such that some fixed predictor has a small cumulative risk, then the cumulative risk of the online algorithm will also be small. There is a slight catch here in that the
learning rate must be chosen a priori, and the optimal setting depends on . The longer
the sequence of examples, the smaller learning rate we want. We can avoid this by using a
learning rate that starts from a fairly large value and decreases as learning progresses. This
leads to a bound similar to Theorem 1 but with somewhat worse constant coefficients.
`
K# K J K W b
1
# ' ) ( for all
.
l*m
h
! _ 1 ( . Then for any such
' ' W
W
a X
. a X
.5ghf g h ( ` 1 ( g _
, ' b
, L
)
V (23)
b
1
1
Let us now consider
of Theorem 2 to a situation in which we assume that
J theareimplications
the examples
i.i.d. according to some fixed distribution R .
kG , such
' ) ( holds with
W
Theorem 3 Let R be a distribution over
that
#
W
_
_
b
probability for KJ
R . Let ! ` W b < 1 1 where is the
-th online
J 1 that
hypothesis based on an example sequence
is drawn i.i.d. according to R .
l*m , and use at update
the learning rate
Fix
! _
h
1 ( . Then for any such
' ' we have
that '
X , W .. ' X , . g hB g )
h ( ` < 1 ( g ` < 1 V
(24)
0,
Theorem 2 Let
be an example sequence such that
, and use at update the learning rate
Fix
that '
we have
If we know in advance how many examples we are going to draw, we can use a fixed
learning rate as in Theorem 1 and obtain somewhat better constants.
5 Experiments and Discussion
In our experiments we studied the performance of online -SVM algorithms in various
settings. They always yielded competitive performance. Due to space constraints we only
report the findings in novelty detection as given in Figure 1 (where the training algorithm
was fed the patterns sans class labels).
_ _
! m Vm _
Already after one pass through the USPS database (5000 training patterns, 2000 test patterns, each of them of size
pixels), which took in MATLAB less than 15s on a
433MHz Celeron, the results can be used for weeding out badly written digits. The ) to allow for a fixed fraction of detected ?outliers.? Based
setting was used (with
.
on the theoretical analysis of Section 4 we used a decreasing learning rate with
h
<
Conclusion We have presented a range of simple online kernel-based algorithms for a
variety of standard machine learning tasks. The algorithms have constant memory requirements and are computationally cheap at each update step. They allow the ready application
of powerful kernel based methods such as novelty detection to online and time-varying
problems.
Results after one pass through the USPS
database. We used Gaussian RBF kernels
with width
. The learn
ing rate was adjusted to
where
is
the number of iterations. Top: the first 50
patterns which incurred a margin error; bottom left: the 50 worst patterns according to
on the training set, bottom right:
the 50 worst patterns on an unseen test set.
(0! m V ! _
1W
`
o8n
Figure 1: Online novelty detection on the USPS dataset with
! m Vm _ .
Acknowledgments A.S. was supported by the DFG under grant Sm 62/1-1, J.K. &
R.C.W. were supported by the ARC. The authors thank Paul Wankadia for help with the
implementation.
References
[1] K. P. Bennett and O. L. Mangasarian. Robust linear programming discrimination of
two linearly inseparable sets. Optimization Methods and Software, 1:23?34, 1992.
[2] G. Cauwenberghs and T. Poggio. Incremental and decremental support vector machine learning. In T. K. Leen, T. G. Dietterich, and V. Tresp, editors, Advances in
Neural Information Processing Systems 13, pages 409?415. MIT Press, 2001.
[3] L. Csat?o and M. Opper. Sparse representation for gaussian process models. In T. K.
Leen, T. G. Dietterich, and V. Tresp, editors, Advances in Neural Information Processing Systems 13, pages 444?450. MIT Press, 2001.
[4] J. Friedman, T. Hastie, and R. Tibshirani. Additive logistic regression: a statistical
view of boosting. Technical report, Stanford University, Dept. of Statistics, 1998.
[5] C. Gentile. A new approximate maximal margin classification algorithm. In T. K.
Leen, T. G. Dietterich, and V. Tresp, editors, Advances in Neural Information Processing Systems 13, pages 500?506. MIT Press, 2001.
[6] T. Graepel, R. Herbrich, and R. C. Williamson. From margin to sparsity. In T. K. Leen,
T. G. Dietterich, and V. Tresp, editors, Advances in Neural Information Processing
Systems 13, pages 210?216. MIT Press, 2001.
[7] Y. Guo, P. Bartlett, A. Smola, and R. C. Williamson. Norm-based regularization of
boosting. Submitted to Journal of Machine Learning Research, 2001.
[8] M. Herbster. Learning additive models online with fast evaluating kernels. In Proc.
14th Annual Conference on Computational Learning Theory (COLT), pages 444?460.
Springer, 2001.
[9] P. J. Huber. Robust statistics: a review. Annals of Statistics, 43:1041, 1972.
[10] G. S. Kimeldorf and G. Wahba. Some results on Tchebycheffian spline functions.
J. Math. Anal. Applic., 33:82?95, 1971.
[11] J. Kivinen, A.J. Smola, and R.C. Williamson. Large margin classification for moving
targets. Unpublished manuscript, 2001.
[12] Y. Li and P.M. Long. The relaxed online maximum margin algorithm. In S. A. Solla,
T. K. Leen, and K.-R. M?uller, editors, Advances in Neural Information Processing
Systems 12, pages 498?504. MIT Press, 1999.
[13] L. Mason, J. Baxter, P. L. Bartlett, and M. Frean. Functional gradient techniques for
combining hypotheses. In A. J. Smola, P. L. Bartlett, B. Sch?olkopf, and D. Schuurmans, editors, Advances in Large Margin Classifiers, Cambridge, MA, 2000. MIT
Press. 221?246.
[14] B. Sch?olkopf, J. Platt, J. Shawe-Taylor, A. J. Smola, and R. C. Williamson. Estimating
the support of a high-dimensional distribution. Neural Computation, 13(7), 2001.
[15] B. Sch?olkopf, A. Smola, R. C. Williamson, and P. L. Bartlett. New support vector
algorithms. Neural Computation, 12(5):1207?1245, 2000.
[16] V. Vapnik, S. Golowich, and A. Smola. Support vector method for function approximation, regression estimation, and signal processing. In M. Mozer, M. Jordan, and
T. Petsche, editors, Advances in Neural Information Processing Systems 9, pages 281?
287, Cambridge, MA, 1997. MIT Press.
| 2128 |@word version:1 norm:2 closure:1 pick:1 thereby:1 series:1 rkhs:1 comparing:1 yet:1 must:1 written:1 fn:1 additive:2 happen:1 cheap:1 treating:1 update:19 discrimination:1 device:1 parameterization:1 boosting:2 math:1 location:2 herbrich:1 firstly:1 alert:1 become:1 huber:5 roughly:1 themselves:1 decreasing:1 automatically:1 jm:2 cache:1 considering:1 becomes:3 begin:1 estimating:1 moreover:1 underlying:1 bounded:3 kimeldorf:1 kg:2 superlinearly:1 minimizes:1 finding:1 warning:1 guarantee:1 every:2 act:2 classifier:1 platt:1 control:1 grant:1 before:1 engineering:1 limit:1 approximately:1 might:1 plus:1 studied:2 limited:1 range:1 acknowledgment:1 practice:1 digit:1 danger:1 empirical:2 projection:1 convenient:1 word:1 seeing:1 cannot:1 storage:1 risk:10 influence:1 elusive:1 straightforward:2 simplicity:1 estimator:11 gkh:2 analogous:1 annals:1 target:2 controlling:1 programming:1 us:1 hypothesis:6 trick:4 element:1 expensive:1 particularly:2 updating:1 database:2 observed:1 bottom:2 worst:2 solla:1 decrease:1 mozer:1 complexity:2 depend:1 basis:1 usps:3 various:1 regularizer:1 fast:1 detected:1 outcome:2 quite:1 larger:1 stanford:1 say:1 otherwise:8 addi:1 statistic:3 unseen:1 online:16 obviously:1 sequence:5 differentiable:1 took:1 product:1 maximal:1 zm:1 relevant:1 combining:1 date:1 insensitivity:2 intuitive:1 olkopf:3 convergence:1 requirement:2 incremental:3 help:2 derive:2 depending:1 clearer:1 golowich:1 frean:1 school:1 progress:1 eq:1 implies:1 australian:1 correct:1 stochastic:3 shrunk:1 require:1 fix:4 secondly:1 adjusted:2 extension:3 sans:1 hold:3 considered:1 predict:2 inseparable:1 purpose:1 estimation:1 proc:1 label:1 uller:1 mit:7 gaussian:4 always:1 aim:1 rather:2 avoid:4 varying:1 intrusion:1 sense:1 tional:1 nn:5 typically:3 going:1 pixel:1 issue:1 classification:10 arg:1 colt:1 priori:1 fairly:1 once:1 having:1 representer:1 others:2 np:1 report:2 spline:1 njb:1 randomly:1 national:1 dfg:1 qing:1 attempt:1 friedman:1 detection:9 evaluation:1 introduces:1 amenable:1 poggio:1 old:1 kjl:2 penalizes:1 taylor:1 theoretical:3 nentially:1 instance:2 soft:4 wb:1 mhz:1 tg:1 cost:3 deviation:1 predictor:1 successful:1 kn:3 sv:2 adaptively:1 herbster:1 probabilistic:1 vm:2 again:1 squared:2 choose:2 worse:1 derivative:5 return:1 li:1 coefficient:4 depends:2 later:1 view:2 h1:1 analyze:1 cauwenberghs:1 start:1 hf:1 competitive:1 minimize:4 efficiently:2 likewise:1 submitted:1 frequency:1 naturally:1 recovers:1 unsolved:1 con:1 dataset:1 adjusting:1 hilbert:2 graepel:1 manuscript:1 specify:1 leen:5 furthermore:2 smola:7 stage:1 hand:2 receives:1 xzy:1 logistic:2 quality:1 dietterich:4 true:1 kjm:1 regularization:8 equality:1 hence:4 nonzero:1 during:1 width:3 meaning:1 mangasarian:1 recently:2 common:1 functional:5 insensitive:7 jl:9 extend:1 slight:1 cambridge:2 inclusion:1 lcm:1 shawe:1 dot:1 moving:2 longer:1 store:2 seen:1 additional:1 somewhat:2 gentile:1 relaxed:1 converge:2 novelty:8 signal:1 ii:1 desirable:4 exceeds:1 ing:1 technical:1 long:1 prediction:7 variant:1 regression:7 n5:2 iteration:7 kernel:20 addition:2 want:3 decreased:1 grow:3 sch:3 unlike:1 spirit:1 jordan:1 call:1 iii:1 baxter:1 hb:1 variety:1 hastie:1 wahba:1 cn:3 qj:1 whether:1 bartlett:4 trimmed:2 proceed:1 matlab:1 useful:2 tune:1 amount:3 gih:1 svms:1 reduced:2 supplied:1 notice:1 overly:1 per:2 csat:1 tibshirani:1 discrete:1 dropping:2 express:1 four:1 tchebycheffian:1 drawn:2 undertaken:1 utilize:1 asymptotically:1 fraction:3 sum:1 powerful:1 draw:1 bound:9 quadratic:1 yielded:1 badly:1 annual:1 precisely:1 constraint:1 alex:1 expo:1 software:1 argument:1 span:1 min:1 performing:2 thej:1 according:5 combination:1 smaller:2 making:1 outlier:1 restricted:1 taken:1 computationally:3 equation:7 previously:1 discus:1 turn:1 hh:2 needed:2 know:2 fed:1 operation:1 petsche:1 batch:5 top:1 ensure:1 hinge:1 approximating:1 added:1 question:1 already:1 costly:1 md:1 gradient:6 thank:1 trivial:1 length:1 kjc:3 retained:1 minimizing:1 robert:1 gk:1 stated:1 implementation:1 anal:1 unknown:1 upper:3 observation:12 sm:1 arc:1 descent:4 situation:5 reproducing:3 tive:1 pair:1 required:2 unpublished:1 applic:1 able:1 below:2 pattern:7 usually:1 sparsity:1 challenge:1 ofh:1 memory:1 power:1 suitable:1 natural:1 regularized:3 kivinen:2 older:1 ready:1 catch:1 tresp:4 kj:6 prior:1 sg:1 review:1 multiplication:1 loss:30 proven:3 incurred:2 editor:7 supported:2 last:1 free:1 truncation:4 allow:2 perceptron:3 absolute:1 sparse:1 overcome:1 opper:1 evaluating:1 avoids:1 cumulative:5 author:1 commonly:1 adaptive:2 made:1 cope:1 functionals:3 decremental:1 approximate:1 overfitting:1 un:1 learn:1 robust:7 schuurmans:1 expansion:5 williamson:6 complex:2 linearly:3 motivation:1 noise:3 paul:1 allowed:2 canberra:1 wish:1 third:1 theorem:7 specific:2 offset:1 mason:1 svm:1 exists:1 vapnik:1 margin:13 easier:1 forget:1 logarithmic:1 simply:1 likely:1 applies:1 springer:1 determines:1 ma:2 goal:2 jyrki:1 consequently:3 rbf:1 towards:1 bennett:1 change:4 specifically:1 typical:1 total:1 pas:2 zone:4 support:5 guo:1 latter:1 dept:1 |
1,239 | 2,129 | Approximate Dynamic Programming
via Linear Programming
Daniela P. de Farias
Department of Management Science and Engineering
Stanford University
Stanford, CA 94305
pucci @stanford.edu
Benjamin Van Roy
Department of Management Science and Engineering
Stanford University
Stanford, CA 94305
bvr@stanford. edu
Abstract
The curse of dimensionality gives rise to prohibitive computational
requirements that render infeasible the exact solution of large- scale
stochastic control problems. We study an efficient method based
on linear programming for approximating solutions to such problems. The approach "fits" a linear combination of pre- selected
basis functions to the dynamic programming cost- to- go function.
We develop bounds on the approximation error and present experimental results in the domain of queueing network control, providing
empirical support for the methodology.
1
Introduction
Dynamic programming offers a unified approach to solving problems of stochastic
control. Central to the methodology is the cost- to- go function, which can obtained
via solving Bellman's equation. The domain of the cost- to- go function is the state
space of the system to be controlled, and dynamic programming algorithms compute and store a table consisting of one cost- to- go value per state. Unfortunately,
t he size of a state space typically grows exponentially in the number of state variables. Known as the curse of dimensionality, this phenomenon renders dynamic
programming intractable in the face of problems of practical scale.
One approach to dealing with this difficulty is to generate an approximation within
a parameterized class of functions , in a spirit similar to that of statistical regression. The focus of this paper is on linearly parameterized functions: one tries to
approximate the cost- to- go function J* by a linear combination of prespecified basis functions. Note that this scheme depends on two important preconditions for the
development of an effective approximation. First, we need to choose basis functions
that can closely approximate the desired cost-to-go function. In this respect, a suitable choice requires some practical experience or theoretical analysis that provides
rough information on the shape of the function to be approximated. "Regularities"
associated with the function, for example, can guide the choice of representation.
Second, we need an efficient algorithm that computes an appropriate linear combination.
The algorithm we study is based on a linear programming formulation, originally
proposed by Schweitzer and Seidman [5], that generalizes the linear programming
approach to exact dynamic programming, originally introduced by Manne [4]. We
present an error bound that characterizes the quality of approximations produced
by the linear programming approach. The error is characterized in relative terms,
compared against the "best possible" approximation of the optimal cost-to-go function given the selection of basis functions. This is the first such error bound for
any algorithm that approximates cost- to- go functions of general stochastic control
problems by computing weights for arbitrary collections of basis functions.
2
Stochastic control and linear programming
We consider discrete- time stochastic control problems involving a finite state space
lSI = N. For each state XES, there is a finite set of available
actions A x. Taking action a E A x when the current state is x incurs cost 9a(X) .
State transition probabilities Pa(x,y) represent, for each pair (x,y) of states and
each action a E A x, the probability that the next state will be y given that the
current state is x and the current action is a E Ax.
S of cardinality
A policy u is a mapping from states to actions. Given a policy u, the dynamics of
the system follow a Markov chain with transition probabilities Pu( x)(x, y). For each
policy u, we define a transition matrix Pu whose (x,y)th entry is Pu(x)(x,y).
The problem of stochastic control amounts to selection of a policy that optimizes
a given criterion. In this paper, we will employ as an optimality criterion infinitehorizon discounted cost of the form
Ju(x) =E
[~(i9U(Xd lxo =x] ,
where 9u(X) is used as shorthand for 9u(x)(X) and the discount factor a E (0,1)
reflects inter- temporal preferences. Optimality is attained by any policy that is
greedy with respect to the optimal cost-to-go function J*(x) = minu Ju(x) (a policy
u is called greedy with respect to J if TuJ = T J).
Let us define operators Tu and T by TuJ = 9u +aPuJ and T J = minu (9u + aPuJ).
The optimal cost-to-go function solves uniquely Bellman's equation J = T J. Dynamic programming offers a number of approaches to solving this equation; one of
particular relevance to our paper makes use of linear programming, as we will now
discuss. Consider the problem
max clJ
(1)
S.t.
T J;::: J,
where c is a vector with positive components, which we will refer to as staterelevance wei9hts. It can be shown that any feasible J satisfies J :::; J*. It follows
that, for any set of positive weights c, J* is the unique solution to (1).
Note that each constraint (T J)(x) ;::: J(x) is equivalent to a set of constraints
+ a L.YEs Pa(X ,y) J(y) ;::: J(x), Va E A x, so that the optimization problem
(1) can be represented as an LP, which we refer to as the exact LP.
9a(X)
As mentioned in the introduction, state spaces for practical problems are enormous
due to the curse of dimensionality. Consequently, the linear program of interest involves prohibitively large numbers of variables and constraints. The approximation
algorithm we study reduces dramatically the number of variables.
Let us now introduce the linear programming approach to approximate dynamic
programming. Given pre-selected basis functions (Pl, .. . , cPK, define a matrix If> =
[ cPl
cPK ]. With an aim of computing a weight vector f E ~K such that If>f
is a close approximation to J*, one might pose the following optimization problem:
max
s.t.
(2)
c'lf>r
Tlf>r
2::
If>r.
Given a solution f, one might then hope to generate near- optimal decisions by using
a policy that is greedy with respect to If>f.
As with the case of exact dynamic programming, the optimization problem (2) can
be recast as a linear program. We will refer to this problem as the approximate
LP. Note that, though the number of variables is reduced to K, the number of
constraints remains as large as in the exact LP. Fortunately, we expect that most
of the constraints will become irrelevant, and solutions to the linear program can
be approximated efficiently, as demonstrated in [3] .
3
Error Bounds for the Approximate LP
When the optimal cost- to- go function lies within the span of the basis functions,
solution of the approximate LP yields the exact optimal cost-to-go function. Unfortunately, it is difficult in practice to select a set of basis functions that contains
the optimal cost- to- go function within its span. Instead, basis functions must be
based on heuristics and simplified analyses. One can only hope that the span comes
close to the desired cost- to- go function.
For the approximate LP to be useful , it should deliver good approximations when
the cost- to- go function is near the span of selected basis functions. In this section,
we present a bound that ensure desirable results of this kind.
To set the stage for development of an error bound, let us establish some notation.
First, we introduce the weighted norms, defined by
1IJ111 "~ =
'"' ')'(x) IJ(x)l ,
~
xES
IIJlloo "~ =
max ')'(x) IJ(x)l,
xES
for any ')' : S f-t ~+. Note that both norms allow for uneven weighting of errors
across the state space.
We also introduce an operator H, defined by
(HV)(x) = max
aEAz
L Pa(x, y)V(y),
y
for all V : S f-t R For any V , (HV)(x) represents the maximum expected value
of V (y) if the current state is x and y is a random variable representing the next
state. Based on this operator, we define a scalar
kv =
for each V : S
f-t ~.
m,:x V(x) -
V(x)
a(HV)(x) ,
(3)
We interpret the argument V of H as a "Lyapunov function," while we view kv as
a "Lyapunov stability factor," in a sense that we will now explain. In the upcoming
theorem, we will only be concerned with functions V that are positive and that
make kv nonnegative. Also, our error bound for the approximate LP will grow
proportionately with kv, and we therefore want kv to be small. At a minimum, kv
should be finite , which translates to a condition
a(HV)(x) < V(x) ,
"Ix ES.
(4)
If a were equal to 1, this would look like a Lyapunov stability condition: the
maximum expected value (HV)(x) at the next time step must be less than the
current value V(x). In general, a is less than 1, and this introduces some slack in
the condition. Note also that kv becomes smaller as the (HV)(x)'s become small
relative to the V(x)'s. Hence, kv conveys a degree of "stability," with smaller values
representing stronger stability.
We are now ready to state our main result. For any given function V mapping S
to positive reals, we use l/V as shorthand for a function x I-t l/V(x).
Theorem 3.1 {2} Let f be a solution of the approximate LP. Then, for any v E 3rK
such that (<T>v) (x) > 0 for all xES and aH <T>v < <T>v ,
IIJ* - <T>flkc :::; 2k<I>v(c'<T>v) min
IIJ* - <T>rll oo,l/<I>v?
r
(5)
A proof of Theorem 3.1 can be found in the long version of this paper [2].
We highlight some implications of Theorem 3.1. First, the error bound (5) tells
that the the approximation error yielded by the approximate LP is proportional to
the error associated with the best possible approximation relative to a certain norm
11?lll,l/<I>v. Hence we expect that the approximate LP will have reasonable behavior
- if the choice of basis functions is appropriate, the approximate LP should yield a
relatively good approximation to the cost-to-go function , as long as the constants
k<I>v and c' <T>v remain small.
Note that on the left-hand side of (5), we measure the approximation error with the
weighted norm 11?lkc. Recall that the weight vector c appears in objective function
of the approximate LP (2) and must be chosen. In approximating the solution to a
given stochastic control problem, it seems sensible to weight more heavily portions
of the state space that are visited frequently, so that accuracy will be emphasized
in such regions. As discussed in [2], it seems reasonable that the weight vector c
should be chosen to reflect the relative importance of each state.
Finally, note that the Lyapunov function <T>v plays a central role in the bound of
Theorem 3.1. Its choice influences three terms on the right-hand-side of the bound:
1. the error minr
IIJ* - <T>rll oo,l/<I>v;
2. the Lyapunov stability factor k<I>v;
3. the inner product c' <T>v with the state- relevance weights.
An appropriately chosen Lyapunov function should make all three of these terms
relatively small. Furthermore, for the bound to be useful in practical contexts,
these terms should not grow much with problem size. We now illustrate with an
application in queueing problems how a suitable Lyapunov function could be found
and show how these terms scale with problem size.
3.1
Example: A Queueing Network
Consider a single reentrant line with d queues and finite buffers of size B. We assume
that exogenous arrivals occur at queue 1 with probability p < 1/2. The state x E ~d
indicates the number of jobs in each queue. The cost per stage incurred at state x
is given by
the average number of jobs per queue .
As discussed in [2] , under certain stability assumptions we expect that the optimal
cost-to-go function should satisfy
P2 I
Pl I
O J * ()
::::;
x::::; dX x
+ de x + Po,
for some positive scalars Po, Pl and P2 independent of d. We consider a Lyapunov
function V(x) = ~XIX + C for some constant C > 0, which implies
m}n IIJ* -lJ>rll oo,l/V
< IIJ*lloo,l /V
<
P2XlX + Plelx + dpo
max '-----'---..,-----'-x2: O
XiX + dC
Po
< P2 + Pl + C'
and the above bound is independent of the number of queues in the system.
Now let us study kv. We have
a(HV)(x)
C) +
< a [p
(~XIX + 2X 1/
< V(x)
(a+ap:;~:~),
1
+
(1- p)
(~XIX + C) ]
and it is clear that, for C sufficiently large and independent of d, there is a j3
independent of d such that aHV ::::; j3V, and therefore kv ::::; 1 ~ ,6 .
<
1
Finally, let us consider ciV. Discussion presented in [2] suggests that one might want
to choose c so as to reflect the stationary state distribution. We expect that under
some stability assumptions, the tail of the stationary state distribution will have an
(l!;l+l)d
upper bound with geometric decay [1]. Therefore we let c(x) =
plxl, for
some 0 < P < 1. In this case, c is equivalent to the conditional joint distribution of
d independent and identically distributed geometric random variables conditioned
on the event that they are less than B + 1, and we have
clV = E
[~t, xl + C I Xi < B + 1, i = 1, ... , d]
< 2 (1
~2p)2 + 1 ~ P + C,
where Xi, i = 1, .. . , d are identically distributed geometric random variables with
parameter 1 - p. It follows that clV is uniformly bounded over the number of
queues.
This example shows that the terms involved in the error bound (5) are uniformly
bounded both in the number of states in the system and in the number of state
variables, hence the behavior of the approximate LP does not deteriorate as the
problem size increases.
We finally present a numerical experiment to further illustrate the performance of
the approximate LP.
L
=-r
", - 3 / 11.5
Al - 1/11.5
)
~
~
- 4 / 11.5
1A8 - 2 .5/ 11.5
-----':7
IJ.z - 2 / 11.5
IC>I"'-" 'U
J
)
-
"" - 3/ 11.5
-
~5 - 3 / 11.5
-
A 2 - 1/11.5
) 1"4- 3 .1/11.5
l
Figure 1: System for Example 3.2.
Policy
Average Cost
Table 1: Average number of jobs after 50,000,000 simulation steps
3.2
An Eight-Dimensional Queueing Network
We consider a queueing network with eight queues. The system is depicted in Figure
1, with arrival P'i, i = 1,2) and departure (J.Li, i = 1, ... ,8) probabilities indicated.
The state x E ~8 represents the number of jobs in each queue. The cost-per-state
is g(x) = lxi, and the discount factor 0:: is 0.995. Actions a E {O, 1}8 indicate which
queues are being served; ai = 1 iff a job from queue i is being processed. We
consider only non-iddling policies and, at each t ime step, a server processes jobs
from one of its queues exclusively.
We choose c of the form c(x) = (1 - p)8 plxl. The basis functions are chosen to span
all polynomials in x of degree 2; therefore, the approximate LP has 47 variables.
Constraints (T<I>r)(x) 2: (<I>r)(x) for the approximate LP are generated by sampling
5000 states according to the distribution associated with c. Experiments were performed for p = 0.85,0.9 and 0.95, and p = 0.9 yielded the policy with smallest
average cost.
We compared the performance of the policy yielded by the approximate LP (ALP)
with that of first-in-first-out (FIFO), last-buffer-first-serve (LBFS)l and a policy
that serves the longest queue in each server (LONG). The average number of jobs
in the system for each policy was estimated by simulation. Results are shown in
Table 1. The policy generated by the approximate LP performs significantly better
than each of the heuristics, yielding more than 10% improvement over LBFS, the
second best policy. We expect that even better results could be obtained by refining
the choice of basis functions and state-relevance weights.
4
Closing Remarks and Open Issues
In t his paper we studied the linear programming approach to approximate dynamic
programming for stochastic control problems as a means of alleviating the curse of
1 LBFS serves the job that is closest to leaving the system; for example, if there are jobs
in queue 2 and in queue 6, a job from queue 2 is processed since it will leave the system
after going through only one more queue, whereas the job from queue 6 will still have to
go through two more queues. We also choose to assign higher priority to queue 8 than to
queue 3 since queue 8 has higher departure probability.
dimensionality. We provided an error bound based on certain assumptions on the
basis functions. The bounds were shown to be uniformly bounded in the number
of states and state variables in certain queueing problems.
Several questions remain open and are the object of future investigation: Can the
state-relevance weights in the objective function be chosen in some adaptive way?
Can we add robustness to the approximate LP algorithm to account for errors in the
estimation of costs and transition probabilities, i.e., design an alternative LP with
meaningful performance bounds when problem parameters are just known to be in
a certain range? How do our results extend to the average cost case? How do our
results extend to the infinite-state case? How does the quality of the approximate
value function, measure by the weighted L1 norm , translate into actual performance
of the associated greedy policy?
Acknowledgements
This research was supported by NSF CAREER Grant ECS-9985229, by the ONR
under Grant MURI N00014-00-1-0637, and by an IBM Research Fellowship.
References
[1] Bertsimas, D. , Gamarnik, D. & Tsitsiklis, J. , "Performance of Multiclass Markovian
Queueing Networks via Piecewise Linear Lyapunov Functions," submitted to Annals of
Applied Probability, 2000.
[2] de Farias, D.P. & Van Roy, B. , "The Linear Programming Approach to Approximate
Dynamic Programming," submitted to publication, 200l.
[3] de Farias, D.P. & Van Roy, B., "On Constraint Sampling for Approximate Linear
Programming," , submitted to publication , 200l.
[4] Manne, A.S., "Linear Programming and Sequential Decisions," Management Science
6, No.3, pp. 259-267, 1960.
[5] Schweitzer, P. & Seidmann, A. , "Generalized Polynomial Approximations in Markovian
Decision Processes," Journal of Mathematical Analysis and Applications 110, pp. 568582, 1985.
| 2129 |@word version:1 polynomial:2 seems:2 norm:5 stronger:1 open:2 simulation:2 incurs:1 reentrant:1 contains:1 exclusively:1 staterelevance:1 current:5 dx:1 must:3 numerical:1 shape:1 civ:1 stationary:2 greedy:4 prohibitive:1 selected:3 prespecified:1 provides:1 preference:1 mathematical:1 schweitzer:2 become:2 shorthand:2 introduce:3 deteriorate:1 inter:1 expected:2 behavior:2 frequently:1 bellman:2 discounted:1 actual:1 curse:4 lll:1 cardinality:1 becomes:1 provided:1 notation:1 bounded:3 kind:1 unified:1 temporal:1 xd:1 prohibitively:1 control:9 grant:2 positive:5 engineering:2 ap:1 might:3 studied:1 cpl:1 suggests:1 range:1 practical:4 unique:1 practice:1 lf:1 empirical:1 significantly:1 pre:2 close:2 selection:2 operator:3 context:1 influence:1 equivalent:2 demonstrated:1 rll:3 go:18 his:1 stability:7 annals:1 play:1 heavily:1 alleviating:1 exact:6 programming:23 pa:3 roy:3 approximated:2 muri:1 role:1 hv:7 precondition:1 region:1 mentioned:1 benjamin:1 dynamic:12 solving:3 deliver:1 serve:1 basis:14 farias:3 po:3 joint:1 represented:1 effective:1 tell:1 whose:1 heuristic:2 stanford:6 product:1 tu:1 manne:2 iff:1 translate:1 kv:10 regularity:1 requirement:1 leave:1 object:1 oo:3 develop:1 illustrate:2 pose:1 ij:3 job:11 p2:3 solves:1 involves:1 come:1 implies:1 indicate:1 lyapunov:9 closely:1 stochastic:8 alp:1 assign:1 investigation:1 pl:4 sufficiently:1 ic:1 minu:2 mapping:2 smallest:1 estimation:1 visited:1 reflects:1 weighted:3 hope:2 rough:1 aim:1 publication:2 ax:1 focus:1 refining:1 longest:1 improvement:1 indicates:1 sense:1 typically:1 lj:1 going:1 issue:1 development:2 equal:1 sampling:2 represents:2 look:1 future:1 piecewise:1 employ:1 ime:1 consisting:1 interest:1 introduces:1 yielding:1 chain:1 implication:1 experience:1 desired:2 theoretical:1 cpk:2 markovian:2 cost:25 minr:1 entry:1 ju:2 fifo:1 central:2 reflect:2 management:3 choose:4 priority:1 li:1 account:1 de:4 satisfy:1 depends:1 performed:1 try:1 view:1 exogenous:1 characterizes:1 portion:1 accuracy:1 efficiently:1 yield:2 yes:1 produced:1 served:1 ah:1 submitted:3 explain:1 against:1 pp:2 involved:1 conveys:1 associated:4 proof:1 recall:1 dimensionality:4 appears:1 originally:2 attained:1 higher:2 follow:1 methodology:2 formulation:1 though:1 furthermore:1 just:1 stage:2 hand:2 quality:2 indicated:1 grows:1 hence:3 uniquely:1 criterion:2 generalized:1 performs:1 l1:1 gamarnik:1 exponentially:1 extend:2 discussed:2 he:1 approximates:1 tail:1 interpret:1 refer:3 ai:1 closing:1 pu:3 add:1 closest:1 optimizes:1 irrelevant:1 store:1 certain:5 buffer:2 server:2 n00014:1 onr:1 minimum:1 fortunately:1 desirable:1 reduces:1 characterized:1 offer:2 long:3 lxo:1 controlled:1 va:1 j3:1 involving:1 regression:1 represent:1 whereas:1 want:2 fellowship:1 grow:2 leaving:1 appropriately:1 spirit:1 near:2 identically:2 concerned:1 fit:1 inner:1 multiclass:1 translates:1 render:2 queue:21 action:6 remark:1 dramatically:1 useful:2 clear:1 amount:1 discount:2 processed:2 reduced:1 generate:2 lsi:1 nsf:1 estimated:1 per:4 discrete:1 enormous:1 queueing:7 bertsimas:1 parameterized:2 reasonable:2 decision:3 bound:17 nonnegative:1 yielded:3 occur:1 constraint:7 x2:1 argument:1 optimality:2 span:5 min:1 relatively:2 department:2 according:1 combination:3 seidman:1 across:1 smaller:2 remain:2 lp:21 equation:3 remains:1 daniela:1 discus:1 slack:1 clv:2 serf:2 generalizes:1 available:1 eight:2 appropriate:2 alternative:1 robustness:1 lxi:1 ensure:1 establish:1 approximating:2 upcoming:1 objective:2 question:1 sensible:1 bvr:1 providing:1 difficult:1 unfortunately:2 rise:1 design:1 policy:16 upper:1 markov:1 finite:4 dc:1 arbitrary:1 introduced:1 pair:1 departure:2 program:3 recast:1 max:5 suitable:2 event:1 difficulty:1 representing:2 scheme:1 ready:1 geometric:3 acknowledgement:1 relative:4 expect:5 highlight:1 proportional:1 incurred:1 degree:2 clj:1 ibm:1 supported:1 last:1 infeasible:1 tsitsiklis:1 guide:1 allow:1 side:2 face:1 taking:1 van:3 xix:4 distributed:2 seidmann:1 transition:4 computes:1 collection:1 adaptive:1 simplified:1 ec:1 approximate:25 lloo:1 dealing:1 xi:2 table:3 tuj:2 ca:2 career:1 domain:2 main:1 linearly:1 arrival:2 iij:5 xl:1 lie:1 weighting:1 ix:1 proportionately:1 theorem:5 rk:1 emphasized:1 x:4 decay:1 intractable:1 sequential:1 importance:1 conditioned:1 depicted:1 infinitehorizon:1 scalar:2 pucci:1 a8:1 satisfies:1 conditional:1 consequently:1 feasible:1 infinite:1 uniformly:3 called:1 experimental:1 e:1 meaningful:1 select:1 uneven:1 support:1 relevance:4 phenomenon:1 |
1,240 | 213 | Connectionist Architectures for Multi-Speaker Phoneme Recognition
Connectionist Architectures/or Multi-Speaker
Phoneme Recognition
John B. Hampshire n and Alex Waibel
School of Computer Science
Carnegie Mellon University
Pittsburgh, PA 15213-3890
ABSTRACT
We present a number of Time-Delay Neural Network (TDNN) based
architectures for multi-speaker phoneme recognition (/b,d,g/ task). We
use speech of two females and four males to compare the performance
of the various architectures against a baseline recognition rate of 95.9%
for a single IDNN on the six-speaker /b,d,g/ task. This series of modular designs leads to a highly modular multi-network architecture capable
of performing the six-speaker recognition task at the speaker dependent
rate of 98.4%. In addition to its high recognition rate, the so-called
"Meta-Pi" architecture learns - without direct supervision - to recognize the speech of one particular male speaker using internal models
of other male speakers exclusively.
1 INTRODUCTION
References [1,2] have show the Tune-Delay Neural Network to be an effective classifier
of acoustic phonetic speech from individual speakers. The objective of this research has
been to extend the TDNN paradigm to the multi-speaker phoneme recognition task, with
the eventual goal of producing connectionist structures capable of speaker independent
phoneme recognition. In making the transition from single to multi-speaker tasks, we have
focused on modular architectures that perform the over-all recognition task by integrating
a number of smaller task-specific networks.
203
204
Hampshire and Waibel
Table 1: A synopsis of multi-speaker /b,d,g/ recognition results for six TDNN-based
architectures.
Architecture
Type
TDNN
baseline
single net
single net
PSTDNN
Multiple
TDNNs
Modular
TDNN
SID
multi net
Meta-Pi
multi net
1.1
multi net
multi net
Features
.Frequency shift
invariance
? arbitrated
classification
.2-stage training
.2-stage training
.Multiple TDNN
modules
.2-stage training
.Multiple TDNN
modules
.Bayesian MAP learning
.no explicit speaker LD.
Size
(connections)
Recognition Rate
3-speakers 6-speakers
6,233
97.3%
(I-ply) 5,357
(2-ply) 6,947
18,700
96.8%
97.2%
98.6%
18,650
37,400
144,000
97.3%
144,000
95.9%
97.1 %
-
-
96.3%
98.3%
-
98.4%
DATA
The experimental conditions for this research are detailed in [1]. Japanese speech data
from six professional announcers (2 female, 4 male) was sampled for the /b, d, g/
phonemes (approximately 250 training and 250 testing tokens per phoneme, per speaker).
Training for all of the modular architectures followed a general two-stage process: in
the first stage, speaker-dependent modules were trained on speech tokens from specific
individuals; in the second stage, the over-all modular structure was trained with speech
tokens from all speakers.
1.2
RESULTS
Owing to the number of architectures investigated, we present only brief descriptions of
each structure. Additional references are provided for readers interested in more detailed
descriptions of particular architectures. Table 1 summarizes our recognition results for
all of the network architectures described below. We list the type of architecture (single
or multi network), the important features of the design, its over-all size (in terms of
total connections), and its recognition performance on the specified multi-speaker task.
There are two principal multi-speaker tasks: a three male task, and a four male/two
female task: the six speaker task is considerably more difficult than its three speaker
counterpart, owing to the higher acoustic variance of combined male/female speech.
Connectionist Architectures for Multi-Speaker Phoneme Recognition
F
oot]
o-;J
F2
F1
Figure 1: The Frequency Shifting TDNN (FSTDNN) architecture.
2 ARCHITECTURE DESCRIPTIONS
TDNN: The TDNN [1,2] serves as our baseline multi-speaker experiment. Its recognition performance on single speaker speech is typically 98.5% [1,3]. The high acoustic
variance of speech drawn from six speakers - two of whom are female - reduces
the TDNN's performance significantly (95.9%). This indicates that architectures capable
of adjusting to markedly different speakers are necessary for robust multi-speaker and
speake~independentrecognJtion.
FSTDNN: In this design, a frequency shift invariant feature is added to the original
TDNN paradigm. The resulting architecture maps input speech into a first hidden layer
with three frequency ranges roughly corresponding to the three formants Fl - F3 (see
figure 1). Two variations of the basic design have been tested [4]: the first is a "one-ply"
architecture (depicted in the figure), while the second is a ''two-ply'' structure that uses
two plies of input to first hidden layer connections. While the frequency shift invariance
of this architecture has intuitive appeal, the resulting network has a very small number
of unique connections from the input to the first hidden layer (- 30, I-ply). This paucity
of connections has two ramifications. First, it creates a crude replica of the input layer
state in the first hidden layer, as a result, feature detectors that form in the connections
between the input and first hidden layers of the standard TDNN are now formed in the
connections between the first and second hidden layers of the FSTDNN. Second, the
crude input to first hidden layer replication results in some loss of information; thus, the
feature detectors of the FSTDNN operate on what can be viewed as a degraded version of
205
206
Hampsllire and Waibel
3?Way niIrMecI output
-----
W.:h;l,~
;-;-;?? .,..,..:;:
. .? ~
? ??~
?.
?
.. .,.. ............
.
,
.
.
..
......
iI
........
.
...
... .11;. I........
.
.:::.
.~
"I.
Inputla,.
~
........... , ?
.~..........
.r ~~:= .... ~~~:
~~
.~ :~:;~;;;;~~~
!
i: ~~. :~:;~~~;; j ~
.: ':-: ~:::::i;= i
Figure 2: The Multiple TDNN architecture: three identical networks trained with three
different objective functions.
the original input. The resulting over-all structure's recognition performance is typically
worse (-- 97%) than the baseline TDNN for the multi-speaker fb,d,g/ task.
Multiple TDNN: This design employs three TDNNs trained with the MSE, Cross Entropy
[5], and CFM [3] objective functions (see figure 2). The different objective functions
used to train the TDNNs form consistently different internal representations of the speech
signal. We exploit these differing representations by using the (potentially) conflicting
outputs of the three networks to form a global arbitrated classification decision. Taking
the normalized sum of the three networks' outputs constitutes a simple arbitration scheme
that typically reduces the single IDNN error rate by 30%.
[Modular TDNN: In this design, we use the connection strengths ofTDNNs fully trained
on individual speakers to form the initial connections of a larger multi-speaker network.
This resulting network's higher layer connections are retrained [6] to produce the final
multi-speaker network. This technique allows us to integrate speaker-dependent networks
into a larger structure, limiting the over-all training time and network complexity of the
final multi-speaker architecture. The 3-speaker modular TDNN (shown in figures 3 and
4) shows selective response to different tokens of speech. In figure 3, the network responds to a Idl phone with only one sub-network (associated with speaker "MNM"). In
fact, this Idl is spoken by "MNM". In figure 4, the same network responds to a fbI phone
spoken by "MHT' with all sub-networks. This selective response to utterances indicates
that the network is sensitive to utterances that are prototypical for all speakers as well
Connectionist Architectures for Multi-Speaker Phoneme Recognition
----
. . . . - ... .01
Figure 3: 3-speaker Modular TDNN responding to input with one module.
Figure 4: 3-speaker Modular TDNN responding to input with three modules.
as those that are unique to an individual. The recognition rate for the 3-speaker modular
TDNN is comparable to the baseline TDNN rate (97.3%); however, the 6-speaker modular TDNN (not shown) yields a substantially lower recognition rate (96.3%). We attribute
this degraded performance to the manner in which this modular structure integrates its
sub-networks. In particular, the sub-networks are integrated by the connections from the
second hidden to output layers. This scheme uses a very small number of connections to
perform the integrating function. As the number of speakers increases and the acoustic
variance of their speech becomes significant, the connection topology becomes inadequate
for the increasingly complex integration function. Interconnecting the sub-networks between the first and second hidden layers would probably improve performance, but the
improvement would be at the expense of modularity. We tried using a "Connectionist
Glue" enhancement to the 6-speaker network [4], but found that it did not result in a
significant recognition improvement.
Stimulus Identification (SID) network: This network architecture is conceptually very
similar to the Integrated Neural Network (INN) [7]. Figure 5 illustrates the network
in block diagram form. Stimulus specific networks (in this case, multiple TDNNs) are
trained to recognize the speech of an individual. Each of these multiple TDNNs forms
a module in the over-all network. The modules are integrated by a superstructure (itself
a multiple TDNN) trained to recognize the identity of the input stimulus (speaker). The
output activations of the integrating superstructure constitute multiplicative connections
that gate the outputs of the modules in order to form a global classification decision.
207
208
Hampshire and Waibel
Output_.
SdlDuI ?? LD. Not
Figure 5: A block diagram of the Stimulus identification (SID) network, which is very
similar to the Integrated Neural Network (INN) [7].
Reference [8] details the SID network's performance. The major advantages of this
architecture are its high degree of modularity (all modules and the integrating superstructure can be trained independently) and it's high recognition rate (98.3%). It's major
disadvantage is that it has no explicit mechanism for handling new speakers (see [8]).
The Meta-Pi Network: This network architecture is very similar to the SID network.
Figure 6 illustrates the network in action. Stimulus specific networks (in this case, multiple TDNNs) are trained to recognize the speech of an individual. Each of these multiple
TDNNs forms a module in the over-all network. The modules are integrated by a superstructure (itself a multiple TDNN) trained in Bayesian MAP fashion to maximize the
phoneme recognition rate of the over-all structure: the equations governing the error backpropagation through the Meta-Pi superstructure link the global objective function with
the output states of the network's speaker-dependent modules [8]. As with the the SID
network, the output activations of the integrating superstructure constitute multiplicative
connections that gate the outputs of the modules in order to form a global classification
decision. However, as mentioned above, the integrating superstructure is not trained independently from the modules it integrates. While this Bayesian MAP training procedure
is not as modularized as the SID network's training procedure, the resulting recognition
rate is comparable. Additionally, the Meta-Pi network forms very broad representations
of speaker types in order to perform its integration task. Reference [8] shows that the
Meta-Pi superstructure learns - without direct supervision - to perform its integra-
Connectionist Architectures for Multi-Speaker Phoneme Recognition
??
. . '.11""?'" ?...., ,. ...
~
~
??
:.::::::
... :.~' ~
..
...
tit
~:::: :
Input ....,..
.~.::::.~::~::: ~~
':.:=::
n
... ..:~:::::
.....
...........
... ??
"
?????????
_
I
..
Figure 6: The Meta-Pi network responding to the speech of one male (MHT) using
models of other males' speech exclusively.
tion function based on gross formant features of the speakers being processed; explicit
speaker identity is irrelevant. A by-product of this learning procedure and the general
representations that it fonns is that the Meta-Pi network learns to recognize the speech
of one male using modules trained for other males exclusively (see figure 6 and [8]).
3 CONCLUSION
We have presented a number ofTDNN-based connectionist architectures for multi-speaker
phoneme recognition. The Meta-Pi network combines the best features of a number of
these designs with a Bayesian MAP learning rule to fonn a connectionist classifier that
performs multi-speaker phoneme recognition at speaker-dependent rates. We believe that
the Meta-Pi network's ability to recognize the speech of one male using only models
of other male speakers is significant. It suggests speech recognition systems that can
maintain their own database of speaker models, adapting to new speakers when possible,
spawning new speaker-dependent learning processes when necessary, and eliminating
redundant or obsolete speaker-dependent modules when appropriate. The one major
disadvantage of the Meta-Pi network is its size. We are presently attempting to reduce the
network's size by 67% (target size: 48,000 connections) without a statistically significant
loss in recognition performance.
209
210
Hampshire and Waibel
Acknowledgements
We wish to thank Bell Communications Research, ATR Interpreting Telephony Research
Laboratories, and the National Science Foundation (EET-8716324) for their support of
this research. We thank Bellcore's David Burr, Daniel Kahn, and Candace Kamm and
Seimens' Stephen Hanson for their comments and suggestions, all of which served to
improve this work. We also thank CMU's Warp/iWarpl group for their support of our
computational requirements. Finally, we thank Barak Pearlmutter, Dean Pomerleau, and
Roni Rosenfeld for their stimulating conversations, insight, and constructive criticism.
References
[1] Waibel, A., Hanazawa, T., Hinton, G., Shikano, K., and Lang, K., "Phoneme
Recognition Using Time-Delay Neural Networks," IEEE Transactions on Acoustics. Speech and Signal Processing, vol. ASSP-37, March, 1989, pp. 328-339.
[2] Lang, K. "A Time-Delay Neural Network Architecture for Speech Recognition,"
Ph.D. Dissertation, Carnegie Mellon University technical report CMU-CS-89-185,
July, 31, 1989.
[3] Hampshire, J., Waibel, A., "A Novel Objective Function for Improved Phoneme
Recognition Using Time-Delay Neural Networks," Carnegie Mellon University
Technical Report CMU-CS-89-118, March, 1989. A shorter version of this technical report is published in the IEEE Proceedings of the 1989 International Joint
Conference on Neural Networks. vol. 1. pp. 235-241.
[4] Hampshire, J., Waibel, A., "Connectionist Architectures for Multi-Speaker Phoneme
Recognition," Carnegie Mellon University Technical Report CMU-CS-89-167, August, 1989.
[5] Hinton, G. E., "Connectionist Learning Procedures," Carnegie Mellon University
Technical Report CMU-CS-87-115 (version 2), December, 1987, pg. 14.
[6] Waibel, A., Sawai, H., and Shikano, K., "Modularity and Scaling in Large Phonemic
Neural Networks", IEEE Transactions on Acoustics. Speech and Signal Processing,
vol. ASSP-37, December, 1989, pp. 1888-1898.
[7] Matsuoka, T., Hamada, H., and Nakatsu, R., "Syllable Recognition Using Integrated
Neural Networks," IEEE Proceedings of the 1989 International Joint Conference
on Neural Networks, Washington, D.C., June 18-22, 1989, vol. 1, pp. 251-258.
[8] Hampshire, J., Waibel, A., ''The Meta-Pi Network: Building Distributed Knowledge Representations for Robust Pattern Recognition," Carnegie Mellon University
Technical Report CMU-CS-89-166, August, 1989.
liWarp is a registered trademark of Intel Corporation.
| 213 |@word version:3 eliminating:1 glue:1 tried:1 pg:1 idl:2 fonn:1 ld:2 initial:1 series:1 exclusively:3 daniel:1 activation:2 lang:2 john:1 obsolete:1 dissertation:1 direct:2 replication:1 combine:1 burr:1 manner:1 roughly:1 multi:26 formants:1 superstructure:8 kamm:1 becomes:2 provided:1 what:1 substantially:1 differing:1 spoken:2 corporation:1 classifier:2 producing:1 approximately:1 suggests:1 range:1 statistically:1 unique:2 testing:1 block:2 backpropagation:1 procedure:4 oot:1 significantly:1 adapting:1 bell:1 integrating:6 map:5 dean:1 independently:2 focused:1 rule:1 insight:1 variation:1 limiting:1 target:1 us:2 pa:1 recognition:34 database:1 module:16 mentioned:1 gross:1 complexity:1 trained:12 tit:1 creates:1 f2:1 joint:2 various:1 mht:2 train:1 effective:1 modular:13 larger:2 ability:1 formant:1 rosenfeld:1 itself:2 hanazawa:1 final:2 advantage:1 net:6 inn:2 product:1 ramification:1 intuitive:1 description:3 enhancement:1 requirement:1 produce:1 school:1 phonemic:1 c:5 owing:2 attribute:1 f1:1 cfm:1 major:3 integrates:2 sensitive:1 june:1 improvement:2 consistently:1 indicates:2 criticism:1 baseline:5 dependent:7 typically:3 integrated:6 hidden:9 kahn:1 selective:2 interested:1 classification:4 bellcore:1 integration:2 f3:1 washington:1 identical:1 broad:1 constitutes:1 connectionist:11 stimulus:5 report:6 employ:1 recognize:6 national:1 individual:6 maintain:1 arbitrated:2 highly:1 male:13 capable:3 necessary:2 shorter:1 disadvantage:2 sawai:1 delay:5 inadequate:1 considerably:1 combined:1 international:2 worse:1 candace:1 multiplicative:2 tion:1 formed:1 degraded:2 phoneme:16 variance:3 yield:1 conceptually:1 sid:7 bayesian:4 identification:2 served:1 published:1 detector:2 against:1 frequency:5 pp:4 associated:1 sampled:1 adjusting:1 conversation:1 knowledge:1 higher:2 response:2 synopsis:1 improved:1 governing:1 stage:6 matsuoka:1 believe:1 building:1 normalized:1 counterpart:1 laboratory:1 speaker:61 pearlmutter:1 performs:1 interpreting:1 novel:1 extend:1 mellon:6 significant:4 supervision:2 own:1 female:5 irrelevant:1 phone:2 phonetic:1 meta:12 additional:1 paradigm:2 maximize:1 integra:1 july:1 stephen:1 signal:3 ii:1 multiple:11 spawning:1 reduces:2 redundant:1 technical:6 cross:1 basic:1 fonns:1 cmu:6 tdnns:7 addition:1 diagram:2 operate:1 markedly:1 probably:1 comment:1 december:2 architecture:31 topology:1 reduce:1 shift:3 six:6 speech:23 roni:1 constitute:2 action:1 detailed:2 tune:1 ph:1 processed:1 per:2 carnegie:6 vol:4 group:1 four:2 drawn:1 replica:1 sum:1 reader:1 decision:3 summarizes:1 scaling:1 comparable:2 layer:11 fl:1 followed:1 syllable:1 hamada:1 strength:1 alex:1 performing:1 attempting:1 waibel:10 march:2 smaller:1 increasingly:1 making:1 presently:1 invariant:1 equation:1 mechanism:1 serf:1 fbi:1 appropriate:1 professional:1 gate:2 original:2 responding:3 paucity:1 exploit:1 objective:6 added:1 responds:2 link:1 thank:4 atr:1 whom:1 difficult:1 potentially:1 expense:1 design:7 pomerleau:1 perform:4 hinton:2 communication:1 assp:2 retrained:1 august:2 david:1 specified:1 connection:16 hanson:1 acoustic:6 registered:1 conflicting:1 below:1 mnm:2 pattern:1 shifting:1 scheme:2 improve:2 brief:1 tdnn:25 utterance:2 acknowledgement:1 loss:2 fully:1 idnn:2 prototypical:1 telephony:1 suggestion:1 foundation:1 integrate:1 degree:1 pi:12 token:4 warp:1 barak:1 taking:1 modularized:1 distributed:1 transition:1 fb:1 transaction:2 eet:1 global:4 pittsburgh:1 shikano:2 modularity:3 table:2 additionally:1 robust:2 mse:1 investigated:1 japanese:1 complex:1 did:1 intel:1 fashion:1 interconnecting:1 sub:5 explicit:3 wish:1 crude:2 ply:6 learns:3 specific:4 trademark:1 list:1 appeal:1 illustrates:2 entropy:1 depicted:1 stimulating:1 goal:1 viewed:1 identity:2 eventual:1 principal:1 hampshire:7 called:1 total:1 invariance:2 experimental:1 internal:2 support:2 constructive:1 arbitration:1 tested:1 handling:1 |
1,241 | 2,130 | A Bayesian Model Predicts Human Parse
Preference and Reading Times in Sentence
Processing
Srini Narayanan
SRI International and ICSI Berkeley
[email protected]
Daniel Jurafsky
University of Colorado, Boulder
[email protected]
Abstract
Narayanan and Jurafsky (1998) proposed that human language comprehension can be modeled by treating human comprehenders as Bayesian
reasoners, and modeling the comprehension process with Bayesian decision trees. In this paper we extend the Narayanan and Jurafsky model
to make further predictions about reading time given the probability of
difference parses or interpretations, and test the model against reading
time data from a psycholinguistic experiment.
1
Introduction
Narayanan and Jurafsky (1998) proposed that human language comprehension can be modeled by treating human comprehenders as Bayesian reasoners, and modeling the comprehension process with Bayesian decision trees. In this paper, we show that the model accounts for parse-preference and reading time data from a psycholinguistic experiment on
reading time in ambiguous sentences.
Parsing, (generally called ?sentence processing? when we are referring to human parsing),
is the process of building up syntactic interpretations for a sentence from an input sequence
of written or spoken words. Ambiguity is extremely common in parsing problems, and
previous research on human parsing has focused on showing that many factors play a role
in choosing among the possible interpretations of an ambiguous sentence.
We will focus in this paper on a syntactic ambiguity phenomenon which has been repeatedly investigated: the main-verb (MV), reduced relative (RR) local ambiguity (Frazier &
Rayner, 1987; MacDonald, Pearlmutter, & Seidenberg, 1994; McRae, Spivey-Knowlton,
& Tanenhaus, 1998, inter alia) In this ambiguity, a prefix beginning with a noun-phrase and
an ambiguous verb-form could either be continued as a main clause (as in 1a), or turn out
to be a relative clause modifier of the first noun phrase (as in 1b).
1. a. The cop arrested the forger.
b. The cop arrested by the detective was guilty of taking bribes.
Many factors are known to influence human parse preferences. One such factor is the different lexical/morphological frequencies of the simple past and participial forms of the ambiguous verbform (arrested, in this case). Trueswell (1996) found that verbs like searched,
with a frequency-based preference for the simple past form, caused readers to prefer the
main clause interpretation, while verbs like selected, had a preference for a participle reading, and supported the reduced relative interpretation.
The transitivity preference of the verb also plays a role in human syntactic disambiguation.
Some verbs are preferably transitive, where others are preferably intransitive. The reduced
relative interpretation, since it involves a passive structure, requires that the verb be transitive. MacDonald, Pearlmutter, and Seidenberg (1994), Trueswell, Tanenhaus, and Kello
(1994) and other have shown that verbs which are biased toward an intransitive interpretation also bias readers toward a main clause interpretation.
Previous work has shown that a competition-integration model developed by SpiveyKnowlton (1996) could model human parse preference in reading ambiguous sentences
(McRae et al., 1998). While this model does a nice job of accounting for the reading-time
data, it and similar ?constraint-based? models rely on a complex set of feature values and
factor weights which must be set by hand. Narayanan and Jurafsky (1998) proposed an
alternative Bayesian approach for this constraint-combination problem. A Bayesian approach offers a well-understood formalism for defining probabilistic weights, as well as for
combining those weights. Their Bayesian model is based on the probabilistic beam-search
of Jurafsky (1996), in which each interpretation receives a probability, and interpretations
were pruned if they were much worse than the best interpretation. The model predicted
large increases in reading time when unexpected words appeared which were only compatible with a previously-pruned interpretation. The model was thus only able to characterize
very gross timing effects caused by pruning of interpretations.
In this paper we extend this model?s predictions about reading time to other cases where
the best interpretation turns out to be incompatible with incoming words. In particular, we
suggest that any evidence which causes the probability of the best interpretation to drop
below its next competitor will also cause increases in reading time.
2
The Experimental Data
We test our model on the reading time data from McRae et al. (1998), an experiment focusing on the effect of thematic fit on syntactic ambiguity resolution. The thematic role of
noun phrase ?the cop? in the prefix ?The cop arrested? is ambiguous. In the continuation
?The cop arrested the crook?, the cop is the agent. In the continuation ?The cop arrested by
the FBI agent was convicted for smuggling drugs?, the cop is the theme. The probabilistic
relationship between the noun and the head verb (?arrested?) biases the thematic disambiguation decision. For example, ?cop? is a more likely agent for ?arrest?, while ?crook? is
a more likely theme. McRae et al. (1998) showed that this ?thematic fit? between the noun
and verb affected phrase-by-phrase reading times in sentences like the following:
2. a. The cop / arrested by / the detective / was guilty / of taking / bribes.
b. The crook / arrested by / the detective / was guilty / of taking / bribes.
In a series of experiment on 40 verbs, they found that sentences with good agents (like cop
in 2a) caused longer reading times for the phrase the detective than sentences with good
themes (like crook in 2b). Figure 1 shows that at the initial noun phrase, reading time is
lower for good-agent sentences than good-patient sentences. But at the NP after the word
?by?, reading time is lower for good-patient sentences than good-agent sentences. 1
1
In order to control for other influences on timing, McRae et al. (1998) actually report reading
time deltas between a reduced relative and non-reduced relative for. It is these deltas, rather than raw
reading times, that our model attempts to predict.
Increased Reading Times (compared to
control)
70
60
50
40
30
20
10
0
The cop/crook arrested by
Good Agent
the detective
Good Patient
Figure 1: Self-paced reading times (from Figure 6 of McRae et al. (1998))
After introducing our model in the next section, we show that it predicts this cross-over in
reading time; longer reading time for the initial NP in good-patient sentences, but shorter
reading time for the post-?by? NP in good-patient sentences.
3
The Model and the Input Probabilities
In the Narayanan and Jurafsky (1998) model of sentence processing, each interpretation of
an ambiguous sentence is maintained in parallel, and associated with a probability which
can be computed via a Bayesian belief net. The model pruned low-probability parses, and
hence predicted increases in reading time when reading a word which did not fit into any
available parse. The current paper extends the Narayanan and Jurafsky (1998) model?s predictions about reading time. The model now also predicts extended reading time whenever
an input word causes the best interpretation to drop in probability enough to switch in rank
with another interpretation.
The model consists of a set of probabilities expressing constraints on sentence processing,
and a network that represents their independence relations:
Data
P(Agent verb, initial NP)
P(Patient verb, initial NP)
P(Participle verb)
P(SimplePast verb)
P(transitive verb)
P(intransitive verb)
P(RR initial NP, verb-ed, by)
P(RR initial NP, verb-ed, by,the)
P(Agent initial NP, verb-ed, by, the, NP)
P(MC SCFG prefix)
P(RR SCFG prefix)
Source
McRae et al. (1998)
McRae et al. (1998)
British National Corpus counts
British National Corpus counts
TASA corpus counts
TASA corpus counts
McRae et al. (1998) (.8, .2)
McRae et al. (1998) (.875. .125)
McRae et al. (1998) (4.6 average)
SCFG counts from Penn Treebank
SCFG counts from Penn Treebank
The first constraint expresses the probability that the word ?cop?, for example, is an agent,
given that the verb is ?arrested?. The second constraint expresses the probability that it is
a patient. The third and fourth constraints express the probability that the ?-ed? form of
the verb is a participle versus a simple past form (for example P(Participle ?arrest?)=.81).
These were computed from the POS-tagged British National Corpus. Verb transitivity
probabilities were computed by hand-labeling subcategorization of 100 examples of each
verb in the TASA corpus. (for example P(transitive ?entertain?)=.86). Main clause prior
probabilities were computed by using an SCFG with rule probabilities trained on the Penn
Treebank version of the Brown corpus. See Narayanan and Jurafsky (1998) and Jurafsky
(1996) for more details on probability computation.
4
Construction Processing via Bayes nets
Using Belief nets to model human sentence processing allows us to a) quantitatively evaluate the impact of different independence assumptions in a uniform framework, b) directly
model the impact of highly structured linguistic knowledge sources with local conditional
probability tables, while well known algorithms for updating the Belief net (Jensen (1995))
can compute the global impact of new evidence, and c) develop an on-line interpretation
algorithm, where partial input corresponds to partial evidence on the network, and the update algorithm appropriately marginalizes over unobserved nodes. So as evidence comes
in incrementally, different nodes are instantiated and the posterior probability of different
interpretations changes appropriately.
The crucial insight of our Belief net model
is to view specific interpretations as values of
latent variables that render top-down (
) and bottom-up evidence ( ) conditionally
independent (d-separate them (Pearl, 1988)). Thus syntactic, lexical, argument structure,
and other contextual information acts as prior or causal support for an interpretation, while
bottom-up phonological or graphological and other perceptual information acts as likelihood, evidential, or diagnostic support.
To applyour
to on-line disambiguation, we assume that there are a set of interpre
model
tations (
) that are consistent with the input data. At different stages of the
input, we compute the posterior probabilities of the different interpretations given the top
down and bottom-up evidence seen so far. 2
V = examine-ed type_of(Subj) = witness
P(A | v, ty(Subj))
P(Arg|v)
P(T | v, ty(Subj))
P(Tense|v)
Arg
Tense
AND
MV
thm
Tense = past
Sem_fit = Agent
Semantic_fit
AND
Arg = trans
Tense = pp
Sem_fit = Theme
RR
thm
Figure 2: The Belief net that represents lexical and thematic support for the two interpretations.
Figure 2 reintroduces the basic structure of our belief net model from Narayanan and Jurafsky (1998). Our model requires conditional probability distributions specifying the preference of every verb for different argument structures, as well its preference for different
tenses. We also compute the semantic fit between possible fillers in the input and different
conceptual roles of a given predicate. As shown in Figure 2, the and interpretations require the conjunction of specific values corresponding to tense, semantic fit and
argument structure features. Note that only the interpretation requires the transitive
argument structure.
2
In this paper, we will focus on the support from thematic, and syntactic features for the Reduced Relative (RR) and Main Verb (MV) interpretations at different stages of the input for the
examples
we saw earlier. So we will have two interpretations "!$# where %&')( *+,-*).,/10
2$3
4%& )( * + -* . /,06575 .
S
[.14] NP?> NP XP
S
[.48] S?> NP [V ...
[.92] S?> NP ...
VP
NP
#1[]
NP
NP VP
VP
V
the witness examined
The witness
MAIN VERB
Figure 3: The partial syntactic parse trees for the
ing an generating grammar.
MAIN CLAUSE
and the
S
NP
XP
VP
NP
Det
interpretations assum-
REDUCED RELATIVE
S
NP
examined
REDUCED RELATIVE
N
V
XP
Det
The witness examined
VP
N
The witness examined
Figure 4: The Bayes nets for the partial syntactic parse trees
The conditional probability of a construction given top-down syntactic evidence
is relatively simple to compute in an augmented-stochastic-context-free formalism (partial
parse trees shown in Figure 3 and the corresponding bayes net in Figure 4). Recall that
the prior probability gives the conditional probability of the right hand side of a
rule given the left hand side. The Inside/Outside algorithm applied to a fixed parse tree
structure is obtained exactly by casting parsing as a special instance of belief propagation.
The correspondences are straightforward a) the parse tree is interpreted as a belief network.
b)the non-terminal nodes correspond to random variables, the range of the variables being
the non-terminal alphabet, c) the grammar rules define the conditional probabilities linking
parent and child nodes, d) the
nonterminal at the root, as well as the terminals at the
leaves represent conditioning evidence to the network, and e) Conditioning on this evidence
produces exactly the conditional probabilities for each nonterminal node in the parse tree
and the joint probability distribution of the parse. 3
The overall posterior ratio requires propagating the conjunctive impact of syntactic and
lexical/thematic sources on our model. Furthermore, in computing the conjunctive impact of the lexical/thematic and syntactic support to compute and , we use the
NOISY- AND (assumes exception independence) model (Pearl, 1988) for combining conjunctive sources. In the case of the and interpretations. At various points, we
compute the
posterior
support
for
the
different
interpretations
using
the
following equa
tion.
. The first term is
3
One complication is that the the conditional distribution in a parse tree %&!,#"( $ / is not
the product distribution %&! ( $ /-%&!"( $ / (it is the conjunctive distribution). However, it is possible to generalize
the belief propagation equations to admit conjunctive distributions %&!,#"( $ /
3
and %&$" ( % / . The diagnostic (inside) support becomes & &' / 0)(+*-, ./& &0 /1& &2 /-%&0 12 ( ' /
and the causal support becomes 3&' / 0546(879, :;3&!< /1& &!=)/-%&' #= ( </ (details can be found at
http://www.icsi.berkeley.edu/ snarayan/scfg.ps).
the syntactic support
while
the second is the lexical and thematic support for a particular
interpretation (
).
5
Model results
9
8
7
6
5
MV/RR
4
3
2
1
0
NP verbed
Model Good Agent
by
Human Good Agent
the
Model Good Patient
NP
Human Good Patient
Figure 5: Completion data
We tested our model on sentences with the different verbs in McRae et al. (1998). For
each verb, we ran our model on sentences with Good Agents (GA) and Good Patients (GP)
for the initial NP. Our model results are consistent with the on-line disambiguation studies
with human subjects (human performance data from McRae et al. (1998)) and show that a
Bayesian implementation of probabilistic evidence combination accounts for garden-path
disambiguation effects.
Figure 5 shows the first result that pertains to the model predictions of how thematic fit
might influence sentence completion times. Our model shows close correspondence to the
human judgements about whether a specific ambiguous verb was used in the Main Clause
(MV) or reduced relative (RR) constructions. The human and model predictions were conducted at the verb (The crook arrested), by (the crook arrested by), the (the crook arrested
by the) and Agent NP (the crook arrested by the detective). As in McRae et al. (1998)
the data shows that thematic fit clearly influenced the gated sentence completion task. The
probabilistic account further captured the fact that at the by phrase, the posterior probability of producing an RR interpretation increased sharply, thematic fit and other factors
2.5
0.9
0.8
2.1
0.7
0.6
1.5
P(X)
P(MC)/P(RR)
2
1
0.541
0.13
0
The crook/detective
arrested by
(a)
Good Agent initial NP
0.4
0.3
0.7
0.5
0.5
the
0.2
0.1
0.04
detective
0.1
0
The cop arrested by
Good Patient Initial NP
Good Agent Main Clause
the detective
Good Agent RR
Figure 6: a) MV/RR for the ambiguous region showing a flip for the Good Agent (ga) case.
b) P(MV) and P(RR) for the Good Patient and Good Agent cases.
(b)
influenced both the sharpness and the magnitude of the increase.
The second result pertains to on-line reading times. Figure 6 shows how the human reading
time reduction effects (reduced compared to unreduced interpretations) increase for Good
Agents (GA) but decrease for Good Patients in the ambiguous region. This explains the
reading time data in Figure 1. Our model predicts this larger effect from the fact that the
most probable interpretation for the Good Agent case flips from the MV to the RR interpretation in this region. No such flip occurs for the Good Patient (GP) case. In Figure 6(a), we
see that the GP results already have the MV/RR ratio less than one (the RR interpretation
is superior) while a flip occurs for the GA sentences (from the initial state where MV/RR
to the final state where MV/RR
). Figure 6 (b) shows a more detailed view of
the GA sentences showing the crossing point where the flip occurs. This finding is fairly
robust (
of GA examples) and directly predicts reading time difficulties.
6
Conclusion
We have shown that a Bayesian model of human sentence processing is capable of modeling
reading time data from a syntactic disambiguation task. A Bayesian model extends current
constraint-satisfaction models of sentence processing with a principled way to weight and
combine evidence. Bayesian models have not been widely applied in psycholinguistics. To
our knowledge, this is the first study showing a direct correspondence between the time
course of maintaining the best a posteriori interpretation and reading time difficulty.
We are currently exploring how our results on flipping of preferred interpretation could
be combined with Hale (2001)?s proposal that reading time correlates with surprise (a surprising (low probability) word leads to large amounts of probability mass to be pruned) to
arrive at a structured probabilistic account of a wide variety of psycholinguistic data.
References
Frazier, L., & Rayner, K. (1987). Resolution of syntactic category ambiguities: Eye movements in
parsing lexically ambiguous sentences. Journal of Memory and Language, 26, 505?526.
Hale, J. (2001). A probabilistic earley parser as a psycholinguistic model. Proceedings of NAACL2001.
Jensen, F. (1995). Bayesian Networks. Springer-Verlag.
Jurafsky, D. (1996). A probabilistic model of lexical and syntactic access and disambiguation. Cognitive Science, 20, 137?194.
MacDonald, M.C., Pearlmutter, N.J., & Seidenberg, M.(1994). The lexical nature of syntactic ambiguity resolution. Psychological Review, 101, 676-703.
McRae, K., Spivey-Knowlton, M., & Tanenhaus, M. K.(1998). Modeling the effect of thematic fit
(and other constraints) in on-line sentence comprehension. Journal of Memory and Language,38,
283?312.
Narayanan, S., & Jurafsky, D. (1998). Bayesian models of human sentence processing. In COGSCI98, pp. 752?757 Madison, WI. Lawrence Erlbaum.
Pearl, J. (1988). Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference.
Morgan Kaufman, San Mateo, Ca.
Spivey-Knowlton, M.(1996). Integration of visual and linguistic information: Human data and model
simulations. Ph.D. Thesis, University of Rochester, 1996.
Trueswell, J. C.(1996). The role of lexical frequency in syntactic ambiguity resolution. Journal of
Memory and Language, 35, 566-585.
Trueswell, J. C., Tanenhaus, M. K., & Kello, C. (1994). Verb-specific constraints in sentence processing: Separating effects of lexical preference from garden-paths. Journal of Experimental
Pyschology: Learning, Memory and Cognition, 19(3), 528?553.
| 2130 |@word version:1 sri:1 judgement:1 simulation:1 accounting:1 rayner:2 detective:9 reduction:1 initial:11 series:1 daniel:1 prefix:4 past:4 current:2 contextual:1 surprising:1 conjunctive:5 written:1 parsing:6 must:1 treating:2 drop:2 update:1 selected:1 leaf:1 beginning:1 node:5 complication:1 preference:10 forger:1 direct:1 consists:1 combine:1 inside:2 inter:1 examine:1 terminal:3 becomes:2 mass:1 kaufman:1 interpreted:1 developed:1 spoken:1 unobserved:1 finding:1 berkeley:3 preferably:2 every:1 act:2 spivey:3 exactly:2 control:2 assum:1 penn:3 producing:1 understood:1 local:2 timing:2 path:2 might:1 mateo:1 examined:4 specifying:1 jurafsky:14 range:1 equa:1 drug:1 word:8 suggest:1 ga:6 close:1 unreduced:1 context:1 influence:3 trueswell:4 www:1 lexical:10 straightforward:1 focused:1 resolution:4 sharpness:1 rule:3 insight:1 continued:1 construction:3 play:2 colorado:2 parser:1 crossing:1 updating:1 predicts:5 bottom:3 role:5 region:3 morphological:1 decrease:1 movement:1 icsi:2 ran:1 gross:1 principled:1 trained:1 po:1 joint:1 psycholinguistics:1 arrested:17 various:1 alphabet:1 instantiated:1 labeling:1 choosing:1 outside:1 larger:1 widely:1 plausible:1 grammar:2 syntactic:17 gp:3 noisy:1 cop:14 final:1 sequence:1 rr:18 net:9 product:1 combining:2 competition:1 parent:1 p:1 produce:1 generating:1 develop:1 completion:3 propagating:1 nonterminal:2 interpre:1 job:1 c:1 involves:1 predicted:2 come:1 stochastic:1 human:21 explains:1 require:1 srini:1 probable:1 comprehension:5 exploring:1 lawrence:1 cognition:1 predict:1 currently:1 saw:1 clearly:1 rather:1 casting:1 conjunction:1 linguistic:2 focus:2 frazier:2 rank:1 likelihood:1 posteriori:1 inference:1 relation:1 arg:3 among:1 overall:1 noun:6 integration:2 special:1 fairly:1 phonological:1 represents:2 others:1 np:25 report:1 quantitatively:1 intelligent:1 national:3 attempt:1 highly:1 intransitive:3 capable:1 entertain:1 partial:5 shorter:1 tree:9 causal:2 psychological:1 increased:2 formalism:2 modeling:4 earlier:1 instance:1 phrase:8 introducing:1 uniform:1 predicate:1 conducted:1 erlbaum:1 characterize:1 combined:1 referring:1 international:1 probabilistic:9 thesis:1 ambiguity:8 marginalizes:1 worse:1 admit:1 cognitive:1 account:4 lexically:1 caused:3 mv:11 tion:1 view:2 root:1 bayes:3 parallel:1 rochester:1 correspond:1 vp:5 generalize:1 bayesian:15 raw:1 mc:2 evidential:1 influenced:2 psycholinguistic:4 whenever:1 ed:5 against:1 competitor:1 ty:2 frequency:3 pp:2 associated:1 recall:1 knowledge:2 actually:1 focusing:1 earley:1 furthermore:1 stage:2 hand:4 receives:1 parse:13 propagation:2 incrementally:1 building:1 effect:7 brown:1 tense:6 hence:1 tagged:1 semantic:2 conditionally:1 transitivity:2 self:1 ambiguous:11 maintained:1 arrest:2 pearlmutter:3 passive:1 reasoning:1 common:1 superior:1 clause:8 conditioning:2 extend:2 interpretation:39 linking:1 expressing:1 tanenhaus:4 language:5 had:1 access:1 longer:2 posterior:5 showed:1 reintroduces:1 verlag:1 seen:1 captured:1 morgan:1 ing:1 offer:1 cross:1 post:1 impact:5 prediction:5 basic:1 patient:14 represent:1 beam:1 proposal:1 source:4 crucial:1 appropriately:2 biased:1 subject:1 enough:1 switch:1 independence:3 fit:9 variety:1 det:2 whether:1 render:1 cause:3 repeatedly:1 generally:1 detailed:1 amount:1 ph:1 narayanan:10 category:1 reduced:10 continuation:2 http:1 delta:2 diagnostic:2 affected:1 express:3 fourth:1 kello:2 extends:2 reader:2 arrive:1 disambiguation:7 decision:3 prefer:1 incompatible:1 modifier:1 paced:1 correspondence:3 subj:3 reasoner:2 constraint:9 sharply:1 alia:1 argument:4 extremely:1 pruned:4 relatively:1 structured:2 combination:2 wi:1 boulder:1 equation:1 previously:1 turn:2 count:6 participle:4 flip:5 available:1 apply:1 fbi:1 alternative:1 scfg:6 top:3 assumes:1 maintaining:1 madison:1 already:1 occurs:3 flipping:1 separate:1 macdonald:3 separating:1 guilty:3 toward:2 mcrae:15 modeled:2 relationship:1 ratio:2 comprehenders:2 implementation:1 gated:1 defining:1 extended:1 witness:5 head:1 verb:32 thm:2 sentence:31 pearl:3 trans:1 able:1 below:1 appeared:1 reading:34 memory:4 garden:2 belief:9 satisfaction:1 difficulty:2 rely:1 tations:1 eye:1 transitive:5 nice:1 prior:3 review:1 relative:10 par:2 versus:1 agent:22 consistent:2 xp:3 treebank:3 compatible:1 course:1 supported:1 free:1 bias:2 side:2 wide:1 taking:3 knowlton:3 san:1 far:1 correlate:1 pruning:1 preferred:1 global:1 incoming:1 corpus:7 conceptual:1 search:1 latent:1 seidenberg:3 table:1 nature:1 robust:1 ca:1 investigated:1 complex:1 did:1 main:10 child:1 augmented:1 theme:4 thematic:13 perceptual:1 subcategorization:1 third:1 british:3 down:3 specific:4 showing:4 hale:2 jensen:2 evidence:11 magnitude:1 tasa:3 surprise:1 likely:2 crook:10 visual:1 unexpected:1 springer:1 corresponds:1 conditional:7 change:1 called:1 experimental:2 exception:1 searched:1 support:10 pertains:2 filler:1 evaluate:1 tested:1 phenomenon:1 |
1,242 | 2,131 | Dynamic Time-Alignment Kernel in
Support Vector Machine
Hiroshi Shimodaira
School of Information Science,
Japan Advanced Institute of
Science and Technology
[email protected]
Mitsuru Nakai
School of Information Science,
Japan Advanced Institute of
Science and Technology
[email protected]
Ken-ichi Noma
School of Information Science,
Japan Advanced Institute of
Science and Technology
[email protected]
Shigeki Sagayama
Graduate School of Information Science
and Technology,
The University of Tokyo
[email protected]
Abstract
A new class of Support Vector Machine (SVM) that is applicable to sequential-pattern recognition such as speech recognition is
developed by incorporating an idea of non-linear time alignment
into the kernel function. Since the time-alignment operation of
sequential pattern is embedded in the new kernel function, standard SVM training and classification algorithms can be employed
without further modifications. The proposed SVM (DTAK-SVM)
is evaluated in speaker-dependent speech recognition experiments
of hand-segmented phoneme recognition. Preliminary experimental results show comparable recognition performance with hidden
Markov models (HMMs).
1
Introduction
Support Vector Machine (SVM) [1] is one of the latest and most successful statistical
pattern classifier that utilizes a kernel technique [2, 3]. The basic form of SVM
classifier which classifies an input vector x ? Rn is expressed as
g(x) =
N
X
i=1
?i yi ?(xi ) ? ?(x) + b =
N
X
?i yi K(xi , x) + b,
(1)
i=1
0
where ? is a non-linear mapping function ?(x) : R n 7? Rn , (n n0 ), ??? denotes
the inner product operator, xi , yi and ?i are the i-th training sample, its class label,
and its Lagrange multiplier, respectively, K is a kernel function, and b is a bias.
Despite the successful applications of SVM in the field of pattern recognition such
as character recognition and text classification, SVM has not been applied to speech
recognition that much. This is because SVM assumes that each sample is a vector
of fixed dimension, and hence it can not deal with the variable length sequences
directly. Because of this, most of the efforts that have been made so far to apply
SVM to speech recognition employ linear time normalization, where input feature
vector sequences with different lengths are aligned to same length [4]. A variant
of this approach is a hybrid of SVM and HMM (hidden Markov model), in which
HMM works as a pre-processor to feed time-aligned fixed-dimensional vectors to
SVM [5]. Another approach is to utilize probabilistic generative models as a SVM
kernel function. This includes the Fisher kernels [6, 7], and conditional symmetric
independence (CSI) kernels [8], both of which employ HMMs as the generative models. Since HMMs can treat sequential patterns, SVM that employs the generative
models based on HMMs can handle sequential patterns as well.
In contrast to those approaches, our approach is a direct extension of the original
SVM to the case of variable length sequence. The idea is to incorporate the operation of dynamic time alignment into the kernel function itself. Because of this,
the proposed new SVM is called ?Dynamic Time-Alignment Kernel SVM (DTAKSVM)?. Unlike the SVM with Fisher kernel that requires two training stages with
different training criteria, one is for training the generative models and the second
is for training the SVM, the DTAK-SVM uses one training criterion as well as the
original SVM.
2
Dynamic Time-Alignment Kernel
We consider a sequence of vectors X = (x1 , x2 , ? ? ? , x L ), where xi ? Rn , L is the
length of the sequence, and the notation |X| is sometimes used to represent the
length of the sequence instead. For simplification, we at first assume the so-called
linear SVM that does not employ non-linear mapping function ?. In such case, the
kernel operation in (1) is identical to the inner product operation.
2.1
Formulation for linear kernel
Assume that we have two vector sequences X and V . If these two patterns are
equal in length, i.e. |X| = |V | = L, then the inner product between X and V can
be obtained easily as a summation of each inner product between xk and v k for
k = 1, ? ? ? , L:
X ?V
=
L
X
xk ? v k ,
(2)
k=1
and therefore an SVM classifier can be defined as given in (1). On the other hand
in case where the two sequences are different in length, the inner product can not
be calculated directly. Even in such case, however, some sort of inner product like
operation can be defined if we align the lengths of the patterns. To that end, let
?(k), ?(k) be the time-warping functions of normalized time frame k for the pattern
X and V , respectively, and let ??? be the new inner product operator instead of
the original inner product ???. Then the new inner product between the two vector
sequences X and V can be given by
X ?V
=
L
1X
x?(k) ? v ?(k) ,
L
(3)
k=1
where L is a normalized length that can be either |X|, |V | or arbitrary positive
integer.
There would be two possible types of time-warping functions. One is a linear timewarping function and the other is a non-linear time-warping function. The linear
time-warping function takes the form as
?(k) = d(|X|/L)ke,
?(k) = d(|V |/L)ke,
where dxe is the ceiling function which gives the smallest integer that is greater than
or equal to x. As it can be seen from the definition given above, the linear warping
function is not suitable for continuous speech recognition, i.e. frame-synchronous
processing, because the sequence lengths, |X| and |V |, should be known beforehand.
On the other hand, non-linear time warping, or dynamic time warping (DTW) [9] in
other word, enables frame-synchronous processing. Furthermore, the past research
on speech recognition has shown that the recognition performance by the non-linear
time normalization outperforms the one by the linear time normalization. Because
of these reasons, we focus on the non-linear time warping based on DTW.
Though the original DTW uses a distance/distortion measure and finds the optimal path that minimizes the accumulated distance/distortion, the DTW that is
employed for SVM uses inner product or kernel function instead and finds the optimal path that maximizes the accumulated similarity:
X ?V
subject to
=
max
?,?
L
1 X
m(k)x?(k) ? v ?(k) ,
M??
(4)
k=1
1 ? ?(k) ? ?(k + 1) ? |X|, ?(k + 1) ? ?(k) ? Q,
1 ? ?(k) ? ?(k + 1) ? |V |, ?(k + 1) ? ?(k) ? Q,
(5)
where m(k) is a nonnegative (path) weighting coefficient, M?? is a (path) normalizing factor, and Q is a constant constraining the local continuity. In the standard
PL
DTW, the normalizing factor M ?? is given as k=1 m(k), and the weighting coefficients m(k) are chosen so that M?? is independent of the warping functions.
The above optimization problem can be solved efficiently by dynamic programming.
The recursive formula in the dynamic programming employed in the present study
is as follows
(
)
G(i ? 1, j) + Inp(i, j),
G(i, j) = max G(i ? 1, j ? 1) + 2 Inp(i, j),
(6)
G(i, j ? 1) + Inp(i, j),
where Inp(i, j) is the standard inner product between the two vectors corresponding
to point i and j. As a result, we have
X ? V = G(|X|, |V |)/(|X| + |V |).
2.2
(7)
Formulation for non-linear kernel
In the last subsection, a linear kernel, i.e. the inner product, for two vector sequences with different lengths has been formulated in the framework of dynamic
time-warping. With a little constraint, similar formulation is possible for the case
where SVM?s non-linear mapping function ? is applied to the vector sequences. To
that end, ? is restricted to the one having the following form:
?(X) = (?(x1 ), ?(x2 ), ? ? ? , ?(x L )),
(8)
where ? is a non-linear mapping function that is applied to each frame vector x i ,
as given in (1). It should be noted that under the above restriction ? preserves the
original length of sequence at the cost of losing long-term correlations such as the
one between x1 and xL . As a result, a new class of kernel can be defined by using
the extended inner product introduced in the previous section;
Ks (X, V )
=
?(X) ? ?(V )
=
max
?,?
=
max
?,?
1
M??
L
X
(9)
m(k)?(x?(k) ) ? ?(v ?(k) )
(10)
k=1
L
1 X
m(k)K(x?(k) , v?(k) ).
M??
(11)
k=1
We call this new kernel ?dynamic time-alignment kernel (DTAK)?.
2.3
Properties of the dynamic time-alignment kernel
It has not been proven that the proposed function Ks (, ) is really an SVM?s admissible kernel which guarantees the existence of a feature space. This is because that
the mapping function to a feature space is not independent but dependent on the
given vector sequences. Although a class of data-dependent asymmetric kernel for
SVM has been developed in [10], our proposed function is more complicated and
difficult to analyze because the input data is a vector sequence with variable length
and non-linear time normalization is embedded in the function. Instead, what have
been known about the proposed function so far are (1) Ks is symmetric, (2) Ks
satisfies the Cauchy-Schwartz like inequality described bellow:
Proposition 1
Ks (X, V )2 ? Ks (X, X)Ks (V, V )
(12)
Proof For simplification, we assume that normalized length L is fixed, and omit
m(k) and M?? in (11). Using the standard Cauchy-Schwartz inequality, the following inequality holds:
Ks (X, V )
=
?
max
?,?
L
X
L
X
?(x?(k) ) ? ?(v ?(k) ) =
k=1
L
X
?(x?? (k) ) ? ?(v ?? (k) )
(13)
k=1
k ?(x?? (k) ) kk ?(v ?? (k) ) k,
(14)
k=1
where ?? (k), ?? (k) represent the optimal warping functions that maximize the RHS
of (13). On the other hand,
Ks (X, X)
=
max
?,?
L
X
?(x?(k) ) ? ?(x?(k) ) =
k=1
L
X
?(x?+ (k) ) ? ?(x?+ (k) ). (15)
k=1
Because here we assume that ?+ (k), ?+ (k) are the optimal warping functions that
maximize (15), for any warping functions including ? ? (k), the following inequality
holds:
L
L
X
X
Ks (X, X) ?
?(x?? (k) ) ? ?(x?? (k) ) =
k ?(x?? (k) ) k2 .
(16)
k=1
k=1
In the same manner, the following holds:
Ks (V, V )
?
L
X
k=1
?(v ?? (k) ) ? ?(v ?? (k) ) =
L
X
k=1
k ?(v ?? (k) ) k2 .
(17)
Therefore,
Ks (X, X)Ks (V, V ) ? Ks (X, V )2
L
X
?
k ?(x?? (k) ) k
2
k=1
=
L X
L
X
!
L
X
k ?(v ?? (k) ) k
k=1
2
!
L
X
?
k ?(x?? (k) ) kk ?(v ?? (k) ) k
k=1
k ?(x?? (i) ) kk ?(v ?? (j) ) k ? k ?(x?? (j) ) kk ?(v ?? (i) ) k
i=1 j=i+1
2
?0
!2
(18)
3
DTAK-SVM
Using the dynamic time-alignment kernel (DTAK) introduced in the previous section, the discriminant function of SVM for a sequential pattern is expressed as
g(X)
=
N
X
?i yi ?(X (i) ) ? ?(X) + b
(19)
?i yi Ks (X (i) , X) + b,
(20)
i=1
=
N
X
i=1
where X (i) represents the i-th training pattern. As it can be seen from these
expressions, the SVM discriminant function for time sequence has the same form
with the original SVM except for the difference in kernels. It is straightforward to
deduce the learning problem which is given as
N
min
W,b,?i
subject to
X
1
W ?W +C
?i ,
2
i=1
(i)
yi (W ? ?(X ) + b) ? 1 ? ?i ,
?i ? 0, i = 1, ? ? ? , N.
(21)
(22)
Again, since the formulation of learning problem defined above is almost the same
with that for the original SVM, same training algorithms for the original SVM can
be used to solve the problem.
4
Experiments
Speech recognition experiments were carried out to evaluated the classification performance of DTAK-SVM. As our objective is to evaluate the basic performance
of the proposed method, very limited task, hand-segmented phoneme recognition
task in which positions of target patterns in the utterance are known, was chosen.
Continuous speech recognition task that does not require phoneme labeling would
be our next step.
4.1
Experimental conditions
The details of the experimental conditions are given in Table 1. The training
and evaluation samples were collected from the ATR speech database: A-set (5240
Table 1: Experimental conditions
Speaker dependency
Phoneme classes
Speakers
Training samples
Evaluation samples
Signal sampling
Feature values
Kernel type
Experiment-1
Experiment-2
dependent
dependent
6 voiced consonants
5 vowels
5 males
5 males and 5 females
200 samples per phoneme
500 samples per phoneme
2,035 samples in all per 2500 samples in all per
speaker
speaker
12kHz, 10ms frame-shift
13-MFCCs and 13-?MFCCs
kx ?x k2
RBF (radial basis function): K(xi , xj ) = exp(? i ? 2 j )
100
C=0.1
C=1.0
C=10
95
90
# SVs / # training samples [%]
Correct classification rate [%]
100
85
80
75
70
65
60
55
50
C=0.1
C=1.0
C=10.0
80
60
40
20
0
0
2
4
6
8
RBF-sigma
(a) Recognition performance
10
1
2
3
4
5
6
7
8
9
10
RBF-sigma
(b) Number of SVs
Figure 1: Experimental results for Experiment-1 (6 voiced-consonants recognition)
showing (a) correct classification rate and (b) the number of SVs as a function of ?
(the parameter of RBF kernel).
Japanese words in vocabulary). In consonant-recognition task (Experiment-1), only
six voiced-consonants /b,d,g,m,n,N/ were used to save time. The classification task
of those 6 phonemes without using contextual information is considered as a relatively difficult task, whereas the classification of 5 vowels /a,i,u,e,o/ (Experiment-2)
is considered as an easier task.
To apply SVM that is basically formulated as a two-class classifier to the multiclass problem, ?one against the others? type of strategy was chosen. The proposed
DTAK-SVM has been implemented with the publicly available toolkit, SVMTorch
[11].
4.2
Experimental results
Fig. 1 depicts the experimental results for Experiment-1, where average values over
5 speakers are shown. It can be seen in Fig. 1 that the best performance of 95.8%
was achieved at ? = 2.0 and C = 10. Similar results were obtained for Experiment-2
as given in Fig. 2.
100
95
# SVs / # training samples [%]
Correct classification rate [%]
100
90
85
80
75
70
65
60
55
50
80
60
40
20
0
0
2
4
6
8
RBF-sigma
10
1
2
3
4
5
6
7
8
9
10
RBF-sigma
(a) Recognition performance
(b) Number of SVs
Figure 2: Experimental results for Experiment-2 (5 vowels recognition) showing
(a) correct classification rate and (b) the number of SVs as a function of ? (the
parameter of RBF kernel).
Table 2: Recognition performance comparison of DTAK-SVM with HMM. Results
of Experiment-1 for 1 male and 1 female speakers are shown. (numbers represent
correct classification rate [%])
Model
HMM (1 mix.)
HMM (4 mix.)
HMM (8 mix.)
HMM (16 mix.)
DTAK-SVM
# training samples/phoneme
male
female
50
100 200
50
100 200
75.0 69.1 77.1 72.2 65.5 76.6
83.3 84.7 90.9 77.3 76.4 86.4
82.8 87.0 92.4 74.6 79.3 88.5
79.9 85.0 93.2 72.9 78.7 89.8
83.8 85.9 92.1 83.5 81.8 87.7
Next, the classification performance of DTAK-SVM was compared with that of the
state-of-the-art HMM. In order to see the effect of generalization performance on
the size of training data set and model complexity, experiments were carried out
by varying the number of training samples (50, 100, 200), and mixtures (1,4,8,16)
for each state of HMM. The HMM used in this experiment was a 3-states, continuous density, Gaussian-distribution mixtures with diagonal covariances, contextindependent model. HTK [12] was employed for this purpose. The parameters of
DTAK-SVM were fixed to C = 10, ? = 2.0. The results for Experiment-1 with
respect to 1 male and 1 female speakers are given in Table 2.
It can be said from the experimental results that DTAK-SVM shows better classification performance when the number of training samples is 50, while comparable
performance when the number of samples is 200. One might argue that the number
of training samples used in this experiment is not enough at all for HMM to achieve
best performance. But such shortage of training samples occurs often in HMMbased real-world speech recognition, especially when context-dependent models are
employed, which prevents HMM from improving the generalization performance.
5
Conclusions
A novel approach to extend the SVM framework for the sequential-pattern classification problem has been proposed by embedding a dynamic time-alignment operation
into the kernel. Though long-term correlations between the feature vectors are omitted at the cost of achieving frame-synchronous processing for speech recognition, the
proposed DTAK-SVMs demonstrated comparable performance in hand-segmented
phoneme recognition with HMMs. The DTAK-SVM is potentially applicable to
continuous speech recognition with some extension of One-pass search algorithm
[9].
References
[1] V. N. Vapnik, Statistical Learning Theory. Wiley, 1998.
[2] B. Sch?
olkopf, C. J. Burges, and A. J. Smola, eds., Advances in Kernel Methods.
The MIT Press, 1998.
[3] ?Kernel machine website,? 2000. http://www.kernel-machines.org/.
[4] P. Clarkson, ?On the Use of Support Vector Machines for Phonetic Classification,? in ICASSP99, pp. 585?588, 1999.
[5] A. Ganapathiraju and J. Picone, ?Hybrid SVM/HMM architectures for speech
recognition,? in ICSLP2000, 2000.
[6] Tommi S. Jaakkola and David Haussler, ?Exploiting generative models in discriminative classifiers,? in Advances in Neural Information Processing Systems
11 (M. S. Kearns and S. A. Solla and D. A. Cohn, ed.), pp. 487?493, The MIT
Press, 1999.
[7] N. Smith and M. Niranjan, ?Data-dependent Kernels in SVM classification of
speech patterns,? in ICSLP-2000, vol. 1, pp. 297?300, 2000.
[8] C. Watkins, ?Dynamic Alignment Kernels,? in Advances in Large Margin Classifiers (A. J. Smola and P. L. Bartlett and B. Sch?
olkopf and D. Schuurmans,
ed.), ch. 3, pp. 39?50, The MIT Press, 2000.
[9] L. Rabiner and B. Juang, Fundamental of Speech Recognition. Prentice Hall,
1993.
[10] K. Tsuda, ?Support Vector Classifier with Asymmetric Kernel Functions,? in
European Symposium on Artificial Neural Networks (ESANN), pp. 183?188,
1999.
[11] R. Collobert,
?SVMTorch:
A Support Vector Machine for
Large-Scale
Regression
and
Classification
Problems,?
2000.
http://www.idiap.ch/learning/SVMTorch.html.
[12] ?The Hidden Markov Model Toolkit (HTK).? http://htk.eng.cam.ac.uk/.
| 2131 |@word covariance:1 eng:1 past:1 outperforms:1 contextual:1 noma:1 enables:1 n0:1 generative:5 website:1 xk:2 smith:1 org:1 direct:1 symposium:1 manner:1 little:1 classifies:1 notation:1 maximizes:1 what:1 minimizes:1 developed:2 guarantee:1 classifier:7 k2:3 schwartz:2 uk:1 omit:1 positive:1 local:1 treat:1 despite:1 path:4 might:1 k:15 hmms:5 limited:1 graduate:1 recursive:1 pre:1 word:2 inp:4 radial:1 operator:2 prentice:1 context:1 restriction:1 www:2 demonstrated:1 latest:1 straightforward:1 ke:2 haussler:1 embedding:1 handle:1 target:1 programming:2 losing:1 us:3 recognition:27 asymmetric:2 database:1 solved:1 solla:1 csi:1 complexity:1 cam:1 dynamic:13 basis:1 easily:1 hiroshi:1 artificial:1 labeling:1 solve:1 distortion:2 itself:1 sequence:16 product:13 aligned:2 achieve:1 olkopf:2 exploiting:1 juang:1 ac:5 school:4 sim:1 esann:1 implemented:1 idiap:1 tommi:1 tokyo:2 correct:5 require:1 icslp:1 generalization:2 really:1 preliminary:1 proposition:1 summation:1 extension:2 pl:1 hold:3 considered:2 hall:1 exp:1 mapping:5 smallest:1 omitted:1 purpose:1 applicable:2 label:1 mit:4 gaussian:1 hil:1 varying:1 jaakkola:1 focus:1 contrast:1 dependent:7 accumulated:2 picone:1 hidden:3 classification:16 html:1 art:1 field:1 equal:2 having:1 sampling:1 identical:1 represents:1 others:1 employ:4 preserve:1 vowel:3 evaluation:2 alignment:11 male:5 mixture:2 beforehand:1 tsuda:1 cost:2 successful:2 dependency:1 density:1 fundamental:1 probabilistic:1 again:1 japan:3 includes:1 coefficient:2 collobert:1 analyze:1 sort:1 complicated:1 voiced:3 publicly:1 phoneme:9 efficiently:1 rabiner:1 basically:1 processor:1 ed:3 definition:1 against:1 pp:5 proof:1 subsection:1 feed:1 htk:3 formulation:4 evaluated:2 though:2 furthermore:1 stage:1 smola:2 correlation:2 hand:6 cohn:1 continuity:1 effect:1 normalized:3 multiplier:1 hence:1 symmetric:2 shigeki:1 deal:1 speaker:8 noted:1 criterion:2 m:1 novel:1 khz:1 jp:4 extend:1 mfccs:2 toolkit:2 similarity:1 deduce:1 align:1 female:4 phonetic:1 inequality:4 yi:6 seen:3 greater:1 employed:5 maximize:2 signal:1 mix:4 segmented:3 long:2 niranjan:1 variant:1 basic:2 regression:1 represent:3 kernel:35 normalization:4 sometimes:1 achieved:1 whereas:1 sch:2 unlike:1 subject:2 integer:2 call:1 constraining:1 enough:1 independence:1 xj:1 architecture:1 inner:13 idea:2 multiclass:1 shift:1 synchronous:3 expression:1 six:1 bartlett:1 effort:1 clarkson:1 speech:15 svs:6 shortage:1 svms:1 ken:1 http:3 per:4 vol:1 ichi:1 achieving:1 utilize:1 ganapathiraju:1 nakai:1 almost:1 utilizes:1 comparable:3 simplification:2 icassp99:1 nonnegative:1 constraint:1 x2:2 min:1 relatively:1 shimodaira:1 character:1 modification:1 restricted:1 ceiling:1 end:2 available:1 operation:6 apply:2 save:1 existence:1 original:8 denotes:1 assumes:1 especially:1 warping:13 objective:1 occurs:1 strategy:1 diagonal:1 said:1 distance:2 atr:1 hmm:13 argue:1 cauchy:2 discriminant:2 collected:1 reason:1 length:15 kk:4 difficult:2 potentially:1 sigma:4 markov:3 extended:1 frame:6 rn:3 timewarping:1 arbitrary:1 introduced:2 david:1 pattern:14 max:6 including:1 suitable:1 hybrid:2 advanced:3 technology:4 jaist:3 dtw:5 carried:2 utterance:1 text:1 embedded:2 proven:1 last:1 bias:1 burges:1 institute:3 dimension:1 calculated:1 dxe:1 vocabulary:1 hmmbased:1 world:1 made:1 far:2 consonant:4 xi:5 discriminative:1 continuous:4 search:1 table:4 improving:1 schuurmans:1 japanese:1 european:1 sagayama:2 rh:1 x1:3 fig:3 depicts:1 contextindependent:1 wiley:1 position:1 xl:1 watkins:1 weighting:2 admissible:1 formula:1 showing:2 svm:45 normalizing:2 incorporating:1 vapnik:1 sequential:6 kx:1 margin:1 easier:1 dtak:14 lagrange:1 expressed:2 prevents:1 ch:2 satisfies:1 conditional:1 formulated:2 rbf:7 fisher:2 except:1 bellow:1 kearns:1 called:2 pas:1 experimental:9 support:6 svmtorch:3 incorporate:1 evaluate:1 |
1,243 | 2,132 | Duality, Geometry, and Support Vector
Regression
Jinbo Bi and Kristin P. Bennett
Department of Mathematical Sciences
Rensselaer Polytechnic Institute
Troy, NY 12180
[email protected], [email protected]
Abstract
We develop an intuitive geometric framework for support vector
regression (SVR). By examining when -tubes exist, we show that
SVR can be regarded as a classification problem in the dual space.
Hard and soft -tubes are constructed by separating the convex
or reduced convex hulls respectively of the training data with the
response variable shifted up and down by . A novel SVR model is
proposed based on choosing the max-margin plane between the two
shifted datasets. Maximizing the margin corresponds to shrinking
the effective -tube. In the proposed approach the effects of the
choices of all parameters become clear geometrically.
1
Introduction
Support Vector Machines (SVMs) [6] are a very robust methodology for inference
with minimal parameter choices. Intuitive geometric formulations exist for the
classification case addressing both the error metric and capacity control [1, 2]. For
linearly separable classification, the primal SVM finds the separating plane with
maximum hard margin between two sets. The equivalent dual SVM computes the
closest points in the convex hulls of the data from each class. For the inseparable
case, the primal SVM optimizes the soft margin of separation between the two
classes. The corresponding dual SVM finds the closest points in the reduced convex
hulls. In this paper, we derive analogous arguments for SVM regression (SVR).
We provide a geometric explanation for SVR with the -insensitive loss function.
From the primal perspective, a linear function with no residuals greater than corresponds to an -tube constructed about the data in the space of the data attributes
and the response variable [6] (see e.g. Figure 1(a)). The primary contribution of this
work is a novel geometric interpretation of SVR from the dual perspective along
with a mathematically rigorous derivation of the geometric concepts. In Section
2, for a fixed > 0 we examine the question ?When does a ?perfect? or ?hard?
-tube exist??. With duality analysis, the existence of a hard -tube depends on
the separability of two sets. The two sets consist of the training data augmented
with the response variable shifted up and down by . In the dual space, regression
becomes the classification problem of distinguishing between these two sets. The
geometric formulations developed for the classification case [1] become applicable to
the regression case. We call the resulting formulation convex SVR (C-SVR) since it
is based on convex hulls of the augmented training data. Much like in SVM classification, to compute a hard -tube, C-SVR computes the nearest points in the convex
hulls of the augmented classes. The corresponding maximum margin (max-margin)
planes define the effective -tube. The size of margin determines how much the
effective -tube shrinks. Similarly, to compute a soft -tube, reduced-convex SVR
(RC-SVR) finds the closest points in the reduced convex hulls of the two augmented
sets.
This paper introduces the geometrically intuitive RC-SVR formulation which is a
variation of the classic -SVR [6] and ?-SVR models [5]. If parameters are properly
tuned, the methods perform similarly although not necessarily identically. RCSVR eliminates the pesky parameter C used in -SVR and ?-SVR. The geometric
role or interpretation of C is not known for these formulations. The geometric
roles of the two parameters of RC-SVR, ? and , are very clear, facilitating model
selection, especially for nonexperts. Like ?-SVR, RC-SVR shrinks the -tube and
has a parameter ? controlling the robustness of the solution. The parameter
acts as an upper bound on the size of the allowable -insensitive error function. In
addition, RC-SVR can be solved by fast and scalable nearest-point algorithms such
as those used in [3] for SVM classification.
2
When does a hard -tube exist?
y
y
y
D
?
y
+
D
+
D
?
D
-
x
D-
D-
(b)
(a)
(d)
(c)
x
+
x
x
Figure 1: The (a) primal hard 0-tube, and dual cases: (b) dual strictly separable > 0 ,
(c) dual separable = 0 , and (d) dual inseparable < 0.
SVR constructs a regression model that minimizes some empirical risk measure
regularized to control capacity. Let x be the n predictor variables and y the dependent response variable. In [6], Vapnik proposed using the -insensitive loss function
L (x, y, f) = |y ? f(x)| = max (0, |y ? f(x)| ? ), in which an example is in error if
its residual |y ? f(x)| is greater than . Plotting the points in (x, y) space as in Figure 1(a), we see that for a ?perfect? regression model the data fall in a hard -tube
about the regression line. Let (Xi , yi) be an example where i = 1, 2, ? ? ? , m, Xi is the
ith predictor vector, and yi is its response. The training data are then (X, y) where
Xi is a row of the matrix X ? Rm?n and y ? Rm is the response. A hard -tube
for a fixed > 0 is defined as a plane y = w 0x + b satisfying ?e ? y ? Xw ? be ? e
where e is an m-dimensional vector of ones.
When does a hard -tube exist? Clearly, for large enough such a tube always
exists for finite data. The smallest tube, the 0-tube, can be found by optimizing:
min
w,b,
s.t. ? e ? y ? Xw ? be ? e
(1)
Note that the smallest tube is typically not the -SVR solution. Let D + and D? be
formed by augmenting the data with the response variable respectively increased
and decreased by , i.e. D + = {(Xi , yi + ), i = 1, ? ? ? , m} and D ? = {(Xi , yi ?
), i = 1, ? ? ? , m}. Consider the simple problem in Figure 1(a). For any fixed > 0,
there are three possible cases: > 0 in which strict hard -tubes exist, = 0
in which only 0-tubes exist, and < 0 in which no hard -tubes exist. A strict
hard -tube with no points on the edges of the tube only exists for > 0 . Figure
1(b-d) illustrates what happens in the dual space for each case. The convex hulls of
D+ and D? are drawn along with the max-margin plane in (b) and the supporting
plane in (c) for separating the convex hulls.
Clearly, the existence of the tube is directly related to the separability of D + and
D? . If > 0 then a strict tube exists and the convex hulls of D + and D? are strictly
separable1 . There are infinitely many possible -tubes when > 0 . One can see
that the max-margin plane separating D + and D? corresponds to one such . In
fact this plane forms an ?
tube where > ? ? 0 . If = 0, then the convex hulls
of D+ and D? are separable but not strictly separable. The plane that separates
the two convex hulls forms the 0 tube. In the last case, where < 0 , the two sets
D+ and D? intersect. No -tubes or max-margin planes exist.
It is easy to show by construction that if a hard -tube exists for a given > 0 then
the convex hulls of D + and D? will be separable. If a hard -tube exists, then there
exists (w, b) such that
(y + e) ? Xw ? be ? 0,
(y ? e) ? Xw ? be ? 0.
(2)
X
0
For any convex combination of D + , (y+e)
0 u where e u = 1, u ? 0 of points
(Xi , yi + ), i = 1, 2, ? ? ? , m, we have (y + e)0 u ? w0(X0 u) ? b ? 0. Similarly for
X0
0
D? , (y?e)
0 v where e v = 1, v ? 0 of points (Xi , yi ? ), i = 1, 2, ? ? ? , m, we have
(y ? e)0 v ? w0 (X0 v) ? b ? 0. Then the plane y = w 0x + b in the -tube separates the
two convex hulls. Note the separating plane and the -tube plane are the same. If
no separating plane exists, then there is no tube. Gale?s Theorem2 of the alternative
can be used to precisely characterize the -tube.
0
Theorem 2.1 (Conditions for existence of hard -tube) A hard -tube exists
for a given > 0 if and only if the following system in (u, v) has no solution:
X0 u = X0 v, e0 u = e0 v = 1, (y + e)0 u ? (y ? e)0 v < 0, u ? 0, v ? 0.
(3)
Proof A hard -tube exists if and only if System (2) has a solution. By Gale?s
Theorem of the alternative [4], system (2) has a solution if and only if the following
alternative system has no solution: X0 u = X0 v, e0 u = e0 v, (y + e)0 u ? (y ? e)0 v =
?1, u ? 0, v ? 0. Rescaling by ?1 where ? = e0 u = e0 v > 0 yields the result.
1
We use the following definitions of separation of convex sets. Let D + and D? be
nonempty convex sets. A plane H = {x : w 0 x = ?} is said to separate D + and D? if
w0x ? ?, ?x ? D + and w0 x ? ?, ?x ? D ? . H is said to strictly separate D + and D? if
w0x ? ? + ? for x ? D + , and w0x ? ? ? ? for each x ? D ? where ? is a positive scalar.
2
The system Ax ? c has a (or has no) solution if and only if the alternative system
A0 y = 0, c0 y = ?1, y ? 0 has no (or has a) solution.
Note that if ? 0 then (y + e)0 u ? (y ? e)0 v ? 0. for any (u, v) such that
X0 u = X0 v, e0 u = e0 v = 1, u, v ? 0. So as a consequence of this theorem, if
D+ and D? are separable, then a hard -tube exists.
3
Constructing the -tube
For any > 0 infinitely many possible -tubes exist. Which -tube should be used?
The linear program (1) can be solved to find the smallest 0-tube. But this corresponds to just doing empirical risk minimization and may result in poor generalization due to overfitting. We know capacity control or structural risk minimization is
fundamental to the success of SVM classification and regression.
We take our inspiration from SVM classification. In hard-margin SVM classification,
the dual SVM formulation constructs the max-margin plane by finding the two
nearest points in the convex hulls of the two classes. The max-margin plane is
the plane bisecting these two points. We know that the existence of the tube is
linked to the separability of the shifted sets, D + and D? . The key insight is that
the regression problem can be regarded as a classification problem between D + and
D? . The two sets D + and D? defined as in Section 2 both contain the same number
of data points. The only significant difference occurs along the y dimension as the
response variable y is shifted up by in D + and down by in D? . For > 0,
the max-margin separating plane corresponds to a hard ?-tube where > ? ? 0.
The resulting tube is smaller than but not necessarily the smallest tube. Figure
1(b) shows the max-margin plane found for > 0 . Figure 1(a) shows that the
corresponding linear regression function for this simple example turns out to be the
0 tube. As in classification, we will have a hard and soft -tube case. The soft
-tube with ? 0 is used to obtain good generalization when there are outliers.
3.1
The hard -tube case
We now apply the dual convex hull method to constructing the max-margin plane
for our augmented sets D + and D? assuming they are strictly separable, i.e. > 0.
The problem is illustrated in detail in Figure 2. The closest points of D + and D? can
be found by solving the following dual C-SVR quadratic program:
min
u,v
s.t.
1
2
X0
(y+e)0
0
u?
X0
(y?e)0
2
v
(4)
e0 u = 1, e v = 1, u ? 0, v ? 0.
X0
Let the closest points in the convex hulls of D + and D? be c = (y+e)
? and
0 u
X0
d = (y?e)0 v
? respectively. The max-margin separating plane bisects these two
? of the plane is the difference between them, i.e., w
points. The normal (w,
? ?)
? =
0
0
0
?
Xu
??Xv
?, ? = (y + e) u
? ? (y ? e)0 v
?. The threshold, ?b, is the distance from the
?
origin
to the
point
halfway
between the two closest points along the normal: b =
y0 u
? +y0 v
?
? +X0 v
?
0 X0 u
0
?
?
?
w
?
+?
. The separating plane has the equation w
? x+ ?y? b = 0.
2
2
Rescaling this plane yields the regression function.
Dual C-SVR (4) is in the dual space. The corresponding Primal C-SVR is:
Figure 2: The solution ?-tube found by C-SVR can have ? < . Squares are original data.
Dots are in D + . Triangles are in D ? . Support Vectors are circled.
1
2
min
w,?,?,?
2
kwk + 12 ? 2 ? (? ? ?)
(5)
Xw + ?(y + e) ? ?e ? 0
Xw + ?(y ? e) ? ?e ? 0.
s.t.
Dual C-SVR (4) can be derived by taking the Wolfe or Lagrangian dual [4] of primal
C-SVR (5) and simplifying.
We prove that the optimal plane from C-SVR bisects the ? tube. The supporting
planes for class D + and class D? determines the lower and upper edges of the ?-tube
respectively. The support vectors from D + and D? correspond to the points along
the lower and upper edges of the ?-tube. See Figure 2.
Theorem 3.1 (C-SVR constructs ?-tube) Let the max-margin plane obtained
? ?b = 0 where w
by C-SVR
? 0x+
?y?
? = X0 u
? ?X0 v
?, ?? = (y+e)0 u
? ?(y?e)0 v
?, and
0(4) be
w
0
0
0
y
u
?
+y
v
?
v
?
0
?b = w
?
? 0 X u?+X
+?
. If > 0, then the plane y = w x + b corresponds
2
2
to an ?-tube of training data (Xi , yi), i = 1, 2, ? ? ? , m where w = ? w??? , b =
? = ?
??
? ??
2??
?
b
??
and
< .
Proof First, we show ?? > 0. By the Wolfe duality theorem [4], ?
? ? ?? > 0,
since the objective values of (5) and the negative objective value of (4) are equal at
optimality. By complementarity, the closest points are right on the margin planes
? ??
? ? ?? = 0 respectively, so ?
? + e)0 u
w
? 0x + ?y
? = 0 and w
? 0x + ?y
?=w
? 0 X0 u
? + ?(y
? and
?
?+
?
?
0 0
0
?
?
?
?
?
?=w
? Xv
? + ?(y?e) v
?. Hence b = 2 , and w,
? ?, ?
? , and ? satisfy the constraints of
?
?
? ? 0. Then subtract the
problem (5), i.e., Xw+
? ?(y+e)?
?
? e ? 0, Xw+
? ?(y?e)?
?e
? ??
? ??
second inequality from the first inequality: 2?
? + ?? ? 0, that is, ?? ? ??
2 > 0
because > 0 ? 0. Rescale constraints by ??? < 0, and reverse the signs. Let
?
w = ? w??? , then the inequalities become Xw ? y ? e ? ???? e, Xw ? y ? ?e ? ??? e.
Let b =
?
b
,
??
then
?
?
??
= b+
??
? ??
2??
??
? ??
= b ? ??
. Substituting into the previous
??
2??
??
? ??
??
? ??
e?be,
Xw?y
?
?
?
e?be. Denote
2??
2??
and
inequalities yields Xw?y ? ?
?
? ?
? = ? ??
< . These inequalities become Xw + be ? y ? ?e, Xw + be ? y ? ??
e.
2??
Hence the plane y = w 0x + b is in the middle of the ?
< tube.
3.2
The soft -tube case
For < 0 , a hard -tube does not exist. Making large to fit outliers may result
in poor overall accuracy. In soft-margin classification, outliers were handled in the
y
2?^
x
Figure 3: Soft ?-tube found by RC-SVR: left: dual, right: primal space.
dual space by using reduced convex hulls. The same strategy works for soft -tubes,
see Figure 3. Instead of taking the full convex hulls of D + and D? , we reduce the
convex hulls away from the difficult boundary cases. RC-SVR computes the closest
points in the reduced convex hulls
2
X0
X0
1
min
u
?
v
0
0
2
(y+e)
(y?e)
u,v
(6)
0
0
s.t.
e u = 1, e v = 1, 0 ? u ? De, 0 ? v ? De.
Parameter D determines the robustness of the solution by reducing the convex hull.
D limits the influence of any single point. As in ?-SVM, we can parameterize D
1
where m is the number of points. Figure 3 illustrates the case
by ?. Let D = ?m
for m = 6 points, ? = 2/6, and D = 1/2. In this example, every
Pmpoint in the
reduced convex hull must depend on at least two data points since i=1 ui = 1 and
0 ? ui ? 1/2. In general, every point in the reduced convex hull can be written as
the convex combination of at least d1/De = d? ? me. Since these points are exactly
the support vectors and there are two reduced convex hulls, 2 ? d?me is a lower
bound on the number of support vectors in RC-SVR. By choosing ? sufficiently
large, the inseparable case with ? 0 is transformed into a separable case where
once again our nearest-points-in-the-convex-hull-problem is well defined.
As in classification, the dual reduced convex hull problem corresponds to computing
a soft -tube in the primal space. Consider the following soft tube version of the
primal C-SVR (7) which has its Wolfe Dual RC-SVR (6):
2
min
1
2
s.t.
Xw + ?(y + e) ? ?e + ? ? 0, ? ? 0
Xw + ?(y ? e) ? ?e ? ? ? 0, ? ? 0
w,?,?,?,? ,?
kwk + 12 ? 2 ? (? ? ?) + C(e0 ? + e0 ?)
(7)
The results of Theorem 3.1 can be easily extended to soft -tubes.
Theorem 3.2 (RC-SVR constructs soft ?-tube) Let the soft max-margin
? ? ?b = 0 where w
plane obtained by RC-SVR (6) be w
? 0x + ?y
? = X0 u
? ? X0 v
?,
0
0
0
0
0
y
u
?
+y
v
?
X
u
?
+X
v
?
0
0
? If 0 < ? 0, then
?? = (y + e) u
? ? (y ? e) v
?, and ?b =
w
?+
?.
2
2
?
? ?
the plane y = w 0x + b corresponds to a soft ? = ? ??
< -tube of training data
2??
(Xi , yi ), i = 1, 2, ? ? ? , m, i.e., a ?-tube of reduced convex hull of training data where
?
?
? + e)0 u
? ? e)0 v
w = ?w
, b = ?b? and ?
?=w
? 0 X0 u
? + ?(y
?, ?? = w
? 0 X0 v
? + ?(y
?.
??
Notice that the ?
? and ?? determine the planes parallel to the regression plane and
through the closest points in each reduced convex hull of shifted data. In the
inseparable case, these planes are parallel but not necessarily identical to the planes
obtained by the primal RC-SVR (7).
Nonlinear C-SVR and RC-SVR can be achieved by using the usual kernel trick. Let
? by a nonlinear mapping of x such that k(Xi , Xj ) = ?(Xi ) ? ?(Xj ). The objective
function of C-SVR (4) and RC-SVR (6) applied to the mapped data becomes
Pm
Pm Pm
1
((ui ? vi )(uj ? vj )(?(Xi ) ? ?(Xj ) + yi yj )) + 2 i=1 (yi (ui ? vi))
j=1
i=1
2
P
P
P
m
m
m
= 21 i=1 j=1 ((ui ? vi )(uj ? vj )(k(Xi , Xj ) + yi yj )) + 2 i=1 (yi (ui ? vi ))
(8)
The final regression P
model after optimizing C-SVR or RC-SVR with kernels takes
m
ui ? v?i ) k(Xi , x) + ?b, where u
?i = u???i , v?i = v???i , ?? = (?
the form of f(x) = i=1 (?
u?
0
0
(?
u
+?
v
)
y
(?
u
+?
v
)
K(?
u
??
v
)
+
where Kij = k(Xi , Xj ).
v
?)0 y + 2, and the intercept term ?b =
2??
4
2
Computational Results
We illustrate the difference between RC-SVR and -SVR on a toy linear problem3.
Figure 4 depicts the functions constructed by RC-SVR and -SVR for different
values of . For large , -SVR produces undesirable results. RC-SVR constructs the
same function for sufficiently large. Too small can result in poor generalization.
2.5
2.5
2
2
1.5
1.5
? = 0.75
1
0.5
0
? = 0.45
0.5
? = 0.25
0
(a)
? = 0.15
?0.5
?1
? = 0.75, 0.45, 0.25
1
(b)
?0.5
0
1
2
3
4
5
6
?1
0
1
2
3
4
5
6
Figure 4: Regression lines from (a) -SVR and (b) RC-SVR with distinct .
In Table 1, we compare RC-SVR, -SVR and ?-SVR on the Boston Housing problem.
Following the experimental design in [5] we used RBF kernel with 2? 2 = 3.9, C =
500?m for -SVR and ?-SVR, and = 3.0 for RC-SVR. RC-SVR, -SVR, and ?-SVR
are computationally similar for good parameter choices. In -SVR, is fixed. In
RC-SVR, is the maximum allowable tube width. Choosing is critical for -SVR
but less so for RC-SVR. Both RC-SVR and ?-SVR can shrink or grow the tube
according to desired robustness. But ?-SVR has no upper bound.
5
Conclusion and Discussion
By examining when -tubes exist, we showed that in the dual space SVR can be
regarded as a classification problem. Hard and soft -tubes are constructed by separating the convex or reduced convex hulls respectively of the training data with
the response variable shifted up and down by . We proposed RC-SVR based on
choosing the soft max-margin plane between the two shifted datasets. Like ?-SVM,
RC-SVR shrinks the -tube. The max-margin determines how much the tube can
shrink. Domain knowledge can be incorporated into the RC-SVR parameters
3
The data consist of (x, y): (0 0), (1 0.1), (2 0.7), (2.5 0.9), (3 1.1) and (5 2). The
CPLEX 6.6 optimization package was used.
Table 1: Testing Results for Boston Housing, MSE= average of mean squared errors of
25 testing points over 100 trials, STD: standard deviation
RC-SVR
-SVR
?-SVR
2?
MSE
STD
MSE
STD
?
MSE
STD
0.1
37.3
72.3
0
11.2
8.3
0.1
9.6
5.8
0.2
11.2
7.6
1
10.8
8.2
0.2
8.9
7.9
0.3
10.7
7.3
2
9.5
8.2
0.3
9.5
8.3
0.4
9.6
7.4
3
10.3
7.3
0.4
10.8
8.2
0.5
8.9
8.4
4
11.6
5.8
0.5
10.9
8.3
0.6
10.6
9.1
5
13.6
5.8
0.6
11.0
8.4
0.7
11.5
9.3
6
15.6
5.9
0.7
11.2
8.5
0.8
12.5
9.8
7
17.2
5.8
0.8
11.1
8.4
and ?. The parameter C in ?-SVM and -SVR has been eliminated. Computationally, no one method is superior for good parameter choices. RC-SVR alone
has a geometrically intuitive framework that allows users to easily grasp the model
and its parameters. Also, RC-SVR can be solved by fast nearest point algorithms.
Considering regression as a classification problem suggests other interesting SVR
formulations. We can show -SVR is equivalent to finding closest points in a reduced
convex hull problem for certain C, but the equivalent problem utilizes a different
metric in the objective function than RC-SVR. Perhaps other variations would yield
even better formulations.
Acknowledgments
Thanks to referees and Bernhard Sch?
olkopf for suggestions to improve this work.
This work was supported by NSF IRI-9702306, NSF IIS-9979860.
References
[1] K. Bennett and E. Bredensteiner. Duality and Geometry in SVM Classifiers. In
P. Langley, eds., Proc. of Seventeenth Intl. Conf. on Machine Learning, p 57?64,
Morgan Kaufmann, San Francisco, 2000.
[2] D. Crisp and C. Burges. A Geometric Interpretation of ?-SVM Classifiers. In
S. Solla, T. Leen, and K. Muller, eds., Advances in Neural Info. Proc. Sys., Vol
12. p 244?251, MIT Press, Cambridge, MA, 1999.
[3] S.S. Keerthi, S.K. Shevade, C. Bhattacharyya and K.R.K. Murthy, A Fast Iterative Nearest Point Algorithm for Support Vector Machine Classifier Design,
IEEE Transactions on Neural Networks, Vol. 11, pp.124-136, 2000.
[4] O. Mangasarian. Nonlinear Programming. SIAM, Philadelphia, 1994.
[5] B. Sch?
olkopf, P. Bartlett, A. Smola and R. Williamson. Shrinking the Tube:
A New Support Vector Regression Algorithm. In M. Kearns, S. Solla, and D.
Cohn eds., Advances in Neural Info. Proc. Sys., Vol 12, MIT Press, Cambridge,
MA, 1999.
[6] V. Vapnik. The Nature of Statistical Learning Theory. Wiley, New York, 1995.
| 2132 |@word trial:1 version:1 middle:1 c0:1 simplifying:1 tuned:1 bhattacharyya:1 jinbo:1 rpi:2 must:1 written:1 alone:1 plane:38 sys:2 ith:1 mathematical:1 along:5 constructed:4 rc:32 become:4 prove:1 x0:24 examine:1 considering:1 becomes:2 what:1 minimizes:1 developed:1 finding:2 every:2 act:1 exactly:1 rm:2 classifier:3 control:3 positive:1 xv:2 limit:1 consequence:1 suggests:1 bredensteiner:1 bi:1 seventeenth:1 acknowledgment:1 yj:2 testing:2 pesky:1 langley:1 intersect:1 empirical:2 svr:82 undesirable:1 selection:1 risk:3 influence:1 intercept:1 crisp:1 equivalent:3 lagrangian:1 maximizing:1 iri:1 convex:38 insight:1 regarded:3 classic:1 variation:2 analogous:1 controlling:1 construction:1 user:1 programming:1 distinguishing:1 origin:1 complementarity:1 wolfe:3 trick:1 satisfying:1 referee:1 std:4 role:2 solved:3 parameterize:1 solla:2 ui:7 depend:1 solving:1 bisecting:1 triangle:1 easily:2 derivation:1 distinct:1 fast:3 effective:3 choosing:4 final:1 housing:2 bennek:1 intuitive:4 olkopf:2 intl:1 produce:1 perfect:2 derive:1 develop:1 illustrate:1 augmenting:1 rescale:1 nearest:6 attribute:1 hull:30 generalization:3 mathematically:1 strictly:5 sufficiently:2 normal:2 mapping:1 substituting:1 inseparable:4 smallest:4 proc:3 applicable:1 bisects:2 kristin:1 minimization:2 mit:2 clearly:2 always:1 ax:1 derived:1 properly:1 rigorous:1 inference:1 dependent:1 typically:1 a0:1 nonexperts:1 transformed:1 overall:1 classification:16 dual:22 equal:1 construct:5 once:1 eliminated:1 identical:1 geometry:2 cplex:1 keerthi:1 grasp:1 introduces:1 primal:10 edge:3 desired:1 e0:11 minimal:1 increased:1 kij:1 soft:17 addressing:1 deviation:1 predictor:2 examining:2 too:1 characterize:1 thanks:1 fundamental:1 siam:1 again:1 squared:1 tube:77 gale:2 conf:1 rescaling:2 toy:1 de:3 satisfy:1 depends:1 vi:4 doing:1 linked:1 kwk:2 parallel:2 contribution:1 formed:1 square:1 accuracy:1 kaufmann:1 yield:4 correspond:1 murthy:1 ed:3 definition:1 pp:1 proof:2 knowledge:1 methodology:1 response:9 formulation:8 leen:1 shrink:5 just:1 smola:1 shevade:1 cohn:1 nonlinear:3 perhaps:1 effect:1 concept:1 contain:1 hence:2 inspiration:1 illustrated:1 width:1 allowable:2 novel:2 mangasarian:1 superior:1 insensitive:3 interpretation:3 significant:1 cambridge:2 pm:3 similarly:3 dot:1 closest:10 showed:1 perspective:2 optimizing:2 optimizes:1 reverse:1 certain:1 inequality:5 success:1 yi:12 muller:1 morgan:1 greater:2 determine:1 ii:1 full:1 scalable:1 regression:17 metric:2 kernel:3 achieved:1 addition:1 decreased:1 grow:1 sch:2 eliminates:1 strict:3 call:1 structural:1 identically:1 enough:1 easy:1 xj:5 fit:1 reduce:1 handled:1 bartlett:1 york:1 clear:2 svms:1 reduced:14 exist:12 nsf:2 shifted:8 notice:1 sign:1 vol:3 key:1 threshold:1 drawn:1 geometrically:3 halfway:1 package:1 separation:2 utilizes:1 bound:3 quadratic:1 precisely:1 constraint:2 argument:1 min:5 optimality:1 separable:9 department:1 according:1 combination:2 poor:3 smaller:1 separability:3 y0:2 making:1 happens:1 outlier:3 computationally:2 equation:1 turn:1 nonempty:1 know:2 apply:1 polytechnic:1 away:1 alternative:4 robustness:3 existence:4 original:1 xw:16 especially:1 uj:2 objective:4 question:1 occurs:1 strategy:1 primary:1 usual:1 said:2 distance:1 separate:4 mapped:1 separating:10 capacity:3 w0:3 me:2 assuming:1 difficult:1 troy:1 info:2 negative:1 design:2 perform:1 upper:4 datasets:2 finite:1 supporting:2 extended:1 incorporated:1 program:2 max:16 explanation:1 critical:1 regularized:1 residual:2 improve:1 philadelphia:1 geometric:9 circled:1 loss:2 interesting:1 suggestion:1 plotting:1 row:1 supported:1 last:1 burges:1 institute:1 fall:1 taking:2 boundary:1 dimension:1 computes:3 san:1 transaction:1 bernhard:1 overfitting:1 francisco:1 xi:15 rensselaer:1 iterative:1 table:2 nature:1 robust:1 mse:4 williamson:1 necessarily:3 constructing:2 domain:1 vj:2 linearly:1 facilitating:1 xu:1 augmented:5 depicts:1 ny:1 wiley:1 shrinking:2 down:4 theorem:7 svm:16 consist:2 exists:10 vapnik:2 illustrates:2 margin:23 subtract:1 boston:2 infinitely:2 scalar:1 corresponds:8 determines:4 ma:2 rbf:1 bennett:2 hard:25 reducing:1 kearns:1 duality:4 experimental:1 support:9 d1:1 problem3:1 |
1,244 | 2,133 | Covariance Kernels from Bayesian
Generative Models
Matthias Seeger
Institute for Adaptive and Neural Computation
University of Edinburgh
5 Forrest Hill, Edinburgh EH1 2QL
[email protected]
Abstract
We propose the framework of mutual information kernels for
learning covariance kernels, as used in Support Vector machines
and Gaussian process classifiers, from unlabeled task data using
Bayesian techniques. We describe an implementation of this framework which uses variational Bayesian mixtures of factor analyzers
in order to attack classification problems in high-dimensional spaces
where labeled data is sparse, but unlabeled data is abundant.
1
Introduction
Kernel machines, such as Support Vector machines or Gaussian processes, are powerful and frequently used tools for solving statistical learning problems. They are
based on the use of a kernel function which encodes task prior knowledge in a
Bayesian manner. In this paper, we propose the framework of mutual information (MI) kernels for learning covariance kernels from unlabeled task data using
Bayesian techniques. This section introduces terms and concepts. We also discuss
some general ideas for discriminative semi-supervised learning and kernel design
in this context. In section 2, we define the general framework and give examples.
We note that the Fisher kernel [4] is a special case of a MI kernel. MI kernels for
mixture models are discussed in detail. In section 3, we describe an implementation
for a MI kernel for variational Bayesian mixtures of factor analyzers models and
show results of preliminary experiments.
the semi-supervised classification problem, a labeled dataset Dl
{(Xl,tl), ... ,(Xm ,tm)} as well as an unlabeled set Du = {x m+1 ,""Xm+n} are
In
given for training, both i.i.d. drawn from the same unknown distribution, but the
labels for Du cannot be observed. Here, Xi E I~.P and ti E {-1, +1}.1 Typically,
m = IDll is rather small, and n = IDul ?m. Our aim is to fit models to Du in a
Bayesian way, thereby extracting (posterior) information, then use this information
to build a covariance kernel K. Afterwards, K will be plugged into a supervised
kernel machine, which is trained on the labeled data Dl to perform the classification
task.
1
For simplicity, we only discuss binary labels here.
It is important to distinguish very clearly between these two learning scenarios.
For fitting D u , we use Bayesian density estimation. After having chosen a model
family {p(xIOn and a prior distribution P(O) over parameters 0, the posterior
distribution P(OIDu) ex P(DuIO)P(O), where P(DuIO) = rr::~'~l P(xiIO), encodes
all information that Du contains about the latent (i.e. unobserved) parameters 0. 2
The other learning scenario is supervised classification, using a kernel machine.
Such architectures model a smooth latent function y (x) E ~ as a random process,
together with a classification noise model P(tly).3 The covariance kernel K specifies
the prior distribution for this process: namely, a-priori, y(x) is assumed to be a
Gaussian process with zero mean and covariance function K , i.e. K(x(1) , X(2 )) =
E[y(x(1))Y(X(2))]; see e.g. [10] for details. In the following, we use the notation
a = (ai)i = (al' ... ,aI)' for vectors, and A = (ai ,j )i,j for matrices respectively. The
prime denotes transposition. diag a is the matrix with diagonal a and 0 elsewhere.
N(xlJ.t,~) denotes the Gaussian density with mean J.t and covariance matrix ~.
Within the standard discriminative Bayesian classification scenario, unlabeled data
cannot be used. However, it is rather straightforward to modify this scenario by
introducing the concept of conditional priors (see [6]). If we have a discriminant
model family {P(tlx; w
a conditional prior P(w 10) allows to encode prior knowledge and assumptions about how information about P(x) (i.e. about 0) influences
our assumptions about a-priori probabilities over discriminants w. For example,
the P(wIO) could be Occam priors, expressing the intuitive fact that for many
problems, the notion of "simplicity" of a discriminant function depends strongly on
what is known about the input distribution P(x). For a given problem, it is in general not easy to come up with a useful conditional prior. However, once such a prior
is specified, we can in principle use the same powerful techniques for approximate
Bayesian inference that have been developed for supervised discriminative settings.
Semi-supervised techniques that can be seen as employing conditional priors include
co-training [1], feature selection based on clustering [7] and the Fisher kernel [4]. For
a probabilistic kernel technique, P( w 10) is fully specified by a covariance function
K(x(1) , X(2) 10) depending on O. The problem is therefore to find covariance kernels
which (as GP priors) favour discriminants in some sense compatible with what we
have learned about the input distribution P(x).
n,
Kernel techniques can be seen as nonparametric smoothers, based on the (prior)
assumption that if two input points are "similar" (e.g. "close" under some distance),
their labels (and latent outputs y) should be highly correlated. Thus, one generic
way of learning kernels from unlabeled data is to learn a distance between input
points from the information about P( x). A frequently used assumption about
how classification labels may depend on P(x) is the cluster hypothesis: we assume
discriminants whose decision boundaries lie between clusters in P(x) to be a-priori
more likely than such that label clusters inconsistently. A general way of encoding
this hypothesis is to learn a distance from P(x) which is consistent with clusters in
P(x) , i.e. points within the same cluster are closer under this distance than points
from different clusters. We can then try to embed the learned distance d(x(1), X(2))
approximately in an Euclidean space, i.e. learn a mapping ? : X r-+ ?( x) E ~l
such that d(x(1) , X(2)) :=;::j 11 ?(x(1)) - ?(X(2)) II for all pairs from Du. Then, a natural
kernel function would be K(x(1) , X(2)) = exp( - ,BII?(x(1)) - ?(x(2))11 2). In this
paper, however, we follow a simpler approach, by considering a similarity measure
2In practice, computation of P(OIDu) is hardly ever feasible , but powerful approximation techniques can be used.
3 A natural choice for binary classification is to represent the log odds log(P(t =
+1Ix)/P(t = -1Ix)) by y(x) .
which immediately gives rise to a covariance kernel, without having to compute an
approximate Euclidean embedding.
Remark: Our main aim in this paper is to construct kernels that can be learned
from unlabeled data only. In contrast to this, the task of learning a kernel from
labeled data is somewhat simpler and can be approached in the following generic
way: start with a parametric model family {y(x; w)} , with the interpretation that
y(x;w) models the log odds log(P(t = +llx)/P(t = -llx)). Fitting these models
to labeled data D[ , we obtain a posterior P(wIDI) . Now , a natural covariance
kernel for our problem is simply K(x(1),X(2)) = Jy(x(1);w)y(x(2 );w)Q(w)dw,
where (say) Q(w) <X P(wID[)AP(W)l - A (or an approximation thereof). For A = 0,
we obtain the prior covariance kernel for our model, while for larger A the kernel
incorporates more and more posterior information. The kernel proposed in [8] can
be seen as approximation to this approach.
2
Mutual Information Kernels
In this section, we begin by introducing the framework of mutual information kernels. Given a mediator distribution Pm e d (()) over parameters (), we define the joint
distribution Q(x(1) , X(2)) mediated by Pm e d (()) as
Q(x(1) , X(2))
=
J
Pmed (())P(x(1)I())P(x(2)1())d().
(1)
The sample mutual information between x(1) and X(2) under this distribution is
(1)
I(x
(2)
,x
_
Q(X(l) , X(2))
) - log Q(x(1))Q(X(2)) '
(2)
where Q(x) = JQ(x , x)dx. I(x(1) , x(2)) is called the mutual information (MI)
score. In a very concrete sense, it measures the similarity between x(1 ) and X(2) with
respect to the generative process represented by the mediator distribution Pm e d (()):
it is the amount of information they share via the mediator variable () ~ Pm e d (()) .
Note that Q(x(1), X(2)) can be seen as inner product in a space of functions () f-t R,
the features of X(k) being (P(x(k)I()))o, weighted by the distribution Pm e d . 4 X(k) is
represented by its likelihood under all possible models.
Covariance kernels have to satisfy the property of positive definiteness 5 , and the MI
score I does not. However, applying a standard transformation (called exponential
embedding (EE) here), we arrive at
K(x(1) X(2)) =
,
e - (I(x (l) ,x(1))+I(x(2),x(2))) /2+I(x(1),x(2))
=
Q(x(1), X(2))
vQ(x(1) , x(1))Q(X(2), X(2))
(3)
EE becomes familiar if we note that it transforms the standard inner product
x(1)' X(2) into the well-known Radial Basis Function (RBF) kernel 6
(4)
4When comparing X ( l) , X ( 2) via the inner product, we are not interested in correlating
their features uniformly, but rather focus on regions of high volume under Pm e d .
5 K is positive definite if the matrix (K(X(k ll , X(k2?)hl ,k2 is positive definite for every
set {x(1 ), ... , X (K ) } of distinct points.
60ne can show that if j is itself a kernel , and j -+ I< under EE, then 1<(3 is also a
kernel for all (3 > 0 (see e.g. [3]) .
or the weighted inner product x(1)'VX(2) into the squared-exponential kernel (e.g.
[10]). It is easy to show that K in (3) is a valid covariance kernel 7 , and we refer to
it as mutual information (MI) kernel.
Example: Let P(xIO) = N(xIO, (p/2)I) (spherical Gaussian with mean 0),
Prned(O) = N(OIO, aI). Then, the MI kernel K is the RBF kernel (4) with
(3 = 4/(p(4 + pia)). Thus, the RBF kernel is a special case of a MI kernel.
2.1
Mediator distribution. Model-trust scaling.
The mediator distribution Prned(O), motivated earlier in this section, should ideally
encode information about the x generation process, just as the Bayesian posterior
P(OIDu). On the other hand , we need to be able to control the influence that
information from sources such as unlabeled data Du can have on the kernel (relying
too much on such sources results in lack of robustness, see e.g. [6] for details). Here,
we propose model-trust scaling (MTS) , by setting
(5)
Prned varies with A from the (usually vague) prior P(O) (A = 0) towards the sharp
posterior P(OID u) (A = n), rendering the Du information (via the model) more
and more influence upon the kernel K. The concrete effect of MTS on the kernel
depends on the model family.
Example (continued): Again, P(xIO) = N(xIO , (p/2)I) , with a flat prior P(O) = 1
on the mean. Then, P(OIDu) = N(Olx , (p/2n)I), where x = n - 1 L:;~;>~l Xi, and
Prned(O) = N(Olx, (p/2A)I) (after (5)). Thus, the MI kernel is again the RBF kernel
(4) with (3 = 2/(p(2 + A)) . For the more flexible model P(xIO) = N(xIJL , ~),
=
(JL,~) and the conjugate Jeffreys prior, the MI kernel is computed in [5].
?
If the Bayesian analysis is done with conjugate prior-model pairs, the corresponding
MI kernel can be computed easily, and for many of these cases, MTS has a very
simple, analytic form (see [5]). In general, approximation techniques developed for
Bayesian analysis have to be applied. For example, applying the Laplace approximation to the computations on a model with flat prior P(O) = 1 results in the Fisher
kernel [4]8, see e.g. [5]. However, in this paper we favour another approximation
technique (see section 3).
2.2
Mutual Information Kernels for Mixture Models
If we apply the MI kernel framework to mixture models P(x 10, 7T") = Ls 7f sP(x lOs),
we run into a problem. As mentioned in section 1, we would like our kernel at least
partly to encode the cluster hypothesis, i.e. K(x(1), X(2)) should be small if x(1), X(2)
come from different clusters in P(x ),9 but the opposite is true (for not too small
7 Q(x(1 ), X ( 2)) is an inner product (therefore a kernel), for the rest of the argument see
[3], section 5.
8This was essentially observed by the authors of [4] on workshop talks, but has not
been published to our knowledge. The fascinating idea of the Fisher kernel has indeed
been the main motivation and inspiration for this paper.
9This does not mean that we (a-priori) believe they should have different labels, but
only that the label (or better: the latent yO) at one of them should not depend strongly
on yO at the other.
A). To overcome this problem, we generalize Q(x(1), X(2)):
S
Q(X(1),X(2)) =
L
WS1 ,S2
J
P(x(1) IOsJP(X(2) IOs2)Prn ed(O) dO,
(6)
8} , 82=1
where W = (W S1 ,S2)Sl ,S2 is symmetric with nonnegative entries and positive elements
on the diagonal. The MI kernel K is defined as before by (3) , based on the new
Q. If Prned(O,rr) = ITsPrn ed(Os)Prn ed(rr) (which is true for the cases we will be
interested in), we see that the original MI kernel arises as special case WS1,S2 =
EPm ed[7fS17fs2]' Now, by choosing W = diag(Epm ed[7fs])s, we arrive at a MI kernel
K which (typically) behaves as expected w.r.t. cluster separation (see figure 1), but
does not exhibit long-range correlations between joined components. In the present
work, we restrict ourselves to this diagonal mixture kernel. Note that this kernel
can be seen as (normalized) mixture of MI kernels over the component models.
Figure 1: Kernel contours on 2-cluster dataset (A = 5,100,30)
Figure 1 shows contour plots 10 of the diagonal mixture kernel for VB-MoFA (see
section 3), learned on a 500 cases dataset sampled from two Gaussians with equal
covariance (see subsection 3.1). We plot K(a , x) for fixed a (marked by a cross)
against all x , the height between contour lines is 0.1. The left and middle plot
have the lower cluster's centre as a, with A = 5, A = 100 respectively, the right
plot's a lies between the cluster centres, A = 30. The effect of MTS can be seen by
comparing left and middle, note the different sharpness of the slopes towards the
other cluster and the different sizes and shapes of the "high correlation" regions. As
seen on the right, points between clusters have highest correlation with other such
inter-cluster points, a feature that may be very useful for successful discrimination.
3
Experiments with Mixtures of Factor Analyzers
In this section, we describe an implementation of a MI kernel , using variational
Bayesian mixtures of factor analyzers (VB-MoFA) [2] as density models. These are
able to combine local dimensionality reduction (using noisy linear transformations
u -+ x from low-dimensional latent spaces) with good global data fit using mixtures.
VB-MoFA is a variational approximation to Bayesian analysis on these models, able
to deliver the posterior approximations we require for an MI kernel.
We employ the diagonal mixture kernel (see subsection 2.2). Instead of implementing MTS analytically, we compute the VB approximation to the true posterior (i.e.
A = n), then simply apply the scaling to this distribution. Prned(0, rr) factorizes as
required in subsection 2.2. The integrals P(x(1) IOs)p(X(2) IOs)Prn ed(Os) dOs in (6)
J
lOProduced using the first-order approximation (see 3) to the MI kernel. Plots using the
one-step variational approximation (see 3) have a somewhat richer structure.
are not analytically tractable. Our first idea was to approximate them by applying
the VB technique once more, ending up with what we called one-step variational
approximations. Unfortunately, the MI kernel approximation based on these terms
cannot be shown to be positive definite anymore l l ! Thus, in the moment we use a
less elegant and, we feel , less accurate approximation (details can be found in [5])
based on first-order Taylor expansions.
In the remainder of this section we compare the VB-MoFA kernel with the RBF
kernel (4) on two datasets, using a Laplace GP classifier (see [10]). In each case
we sample a training pool, a kernel dataset Du and a test set (mutually exclusive).
The VB-MoFA diagonal mixture kernel is learned on Du. For a given training set
size m, a run consists of sampling a training set Dl and a holdout set Dh (both of
size m) from the training pool, tuning kernel parameters by validation on D h , then
testing on the test set. We use the same Dl, Du for both kernels. For each training
set size, we do L = 30 runs. Results are presented by plotting means and 95% t-test
confidence intervals of test errors over runs.
3.1
Two Gaussian clusters
The dataset is sampled from two 2-d Gaussians with same non-spherical covariance
(see figure 1) , one for each class (the Bayes error is 2.64%) . We use n = 500 points
for D u , a training pool of 100 and a test set of 500 points. The learning curves in
figure 2 show that on this simple toy problem, on which the fitted VB-MoFA model
represents the cluster structure in P(x) almost perfectly, the VB-MoFA MI kernel
outperforms the RBF kernel for samples sizes n :::; 40.
,
_ 0 .225
~
~
02
~
0.15
02
~O. 175
~ 0.175
~
IL
0.15
~O.125
rI
~~~~~---7---~
~~~~~~--~
Training
set size n
I
I
',~~~~---7---.~
, --~--~--~~
Trair>ngsets;zen
Figure 2: Learning curves on 2-cluster dataset. Left: RBF kernel; right: MI kernel
3.2
Handwritten Digits (MNIST): Twos against threes
We report results of preliminary experiments using the subset of twos and threes
of the MNIST Handwritten Digits database 12 . Here, n = IDul = 2000, the training
pool contains 8089, the test set 2000 cases. We employ a VB-MoFA model with 20
components, fitted to Du. We use a very simple baseline (BL) algorithm (see [6],
section 2.3) based on the component densities from the VB-MoFA model13 , which
llThanks to an anonymous reviewer for pointing out this flaw.
12The 28 x 28 images were downsampled to size 8 x 8.
13The estimates P(xls) are obtained by integrating out the parameters (}s using the
variational posterior approximation. The integral is not analytic, and we use a one-step
variational approximation to it .
allows us to assess the "purity" of the component clusters w.r.t. the labels 1 \ this
algorithm is the only one not based on a kernel. Furthermore, we show results for
the one-step variational approximation to the MI kernel 15 (MIOLD ). The learning
curves are shown in figure 3.
to.>
...
1
~
r
II
.....
, II
~ 0.'
?
t? .
! ??
T.-." ,",_ ,
,
II I
j"
i
I ..
!
,,_
...
_,
=?
II I I
T,_ ... _,
j"
iI ..
!
[
Figure 3: Learning curves on MNIST twos/threes. Upper left: RBF kernel; upper
middle: Baseline method; upper right: VB-MoFA MI kernel (first-order approx.) ;
lower left: VB-MoFA MI "kernel" (one-step var. approx.)
The results are disappointing. The fact that the first-order approximation to the
MI kernel performs worse than the one-step variational approximation (although
the latter may fail to be positive definite) , indicates that the former is a poorer
approximation. The latter renders results close to the baseline method, while the
smoothing RBF kernel makes much better use of a growing number of labeled examples 16 This indicates that the conditional prior, as represented by the VB-MoFA
MI kernel, behaves nonsmooth and overrides label information in regions where it
should not. We suspect this problem to be related to the high dimensionality of
the input space, in which case probability densities tend to have a large dynamic
range, and mixture component responsibility estimates tend to behave very nonsmooth. Thus, it seems to be necessary to extend the basic MI kernel framework
by new scaling mechanisms in order to produce a smoother encoding of the prior
assumptions.
14The baseline algorithm is based on the assumption that, given the component index
s, the input point x and the label t are independent. Only the conditional probabilities
P(t ls) are learned, while P(xls) and pes) is obtained from the VB-MoFA model fitted to
unlabeled data only. Thus, success/failure of this method should be closely related to the
degree of purity of the component clusters w.r.t . the labels.
15This is somewhat inconsistent, since we use a kernel function which might not be
positive definite in a context (GP classification) which requires a covariance function.
16Note also that RBF kernel matrices can be evaluated significantly faster than such
using the VB-MoFA MI kernel.
4
Related work. Discussion
The present work is probably most closely related to the Fisher kernel (see subsection 2.1). The arguments concerning mixture models (see subsection 2.2) apply
there as well. Haussler [3] contains a wealth of material about kernel design for discrete objects x. Watkins [9] mentions that expressions like Q in (1) are valid kernels
for discrete x and countable parameter spaces. Very recently we came across [11],
which essentially describes a special case of the diagonal mixture kernel (see subsection 2.2) for Gaussian components with diagonal covariances 17 . The author calls
Q a stochastic equivalence predicate. He is interested in distance learning, does not
apply his method to kernel machines and does not give a Bayesian interpretation.
We have presented a general framework for kernel learning from unlabeled data and
described an approximate implementation using VB-MoFA models. A straightforward application of this technique to high-dimensional real-world data did not prove
successful, and in future work we will explore new ideas for extending the basic MI
kernel framework in order to be able to deal with high-dimensional input spaces.
Acknowledgments
We thank Chris Williams for many inspiring discussions , furthermore Ralf Herbrich, Amos Storkey, Hugo Zaragoza and Neil Lawrence. Matt Beal helped us a
lot with VB-MoFA. The author gratefully acknowledges support through a research
studentship from Microsoft Research Ltd.
References
[1] Avrim Blum and Tom Mitchell. Combining labeled and unlabeled data with CoTraining. In Proceedings of COLT, 1998.
[2] Z. Ghahramani and M. Beal. Variational inference for Bayesian mixtures of factor
analysers. In Advances in NIPS 12. MIT Press, 1999.
[3] David Haussler. Convolution kernels on discrete structures. Technical Report UCSCCRL-99-10 , University of California, Santa Cruz, July 1999.
[4] Tommi S. Jaakkola and David Haussler. Exploiting generative models in discriminative classifiers. In Advances in N eural Information Processing Systems 11, 1998.
[5] Matthias Seeger. Covariance kernels from Bayesian generative models. Technical
report , 2000. Available at http : //yyy . dai . ed. ac . ukr seeger /papers . html.
[6] Matthias Seeger. Learning with labeled and unlabeled data. Technical report, 2000.
Available at http://yyy .dai. ed. ac. ukrseeger/papers .html.
[7] Martin Szummer and Tommi Jaakkola. Partially labeled classification with Markov
random walks. In Advances in NIPS
14. MIT Press, 200l.
[8] Koji Tsuda, Motoaki Kawanabe, Gunnar Ratsch, Soeren Sonnenburg, and KlausRobert Muller. A new discriminative kernel from probabilistic models . In Advances
in NIPS 14. MIT Press, 200l.
[9] Chris Watkins. Dynamic alignment kernels. Technical Report CSD-TR-98-11 , Royal
Holloway, University of London, 1999.
[10] Christopher K.1. Williams and David Barber. Bayesian classification with Gaussian
processes. IEEE Trans. PAMI, 20(12):1342- 1351, 1998.
[11] Peter Yianilos. Metric learning via normal mixtures. Technical report , NEC Research ,
Princeton, 1995.
17The a parameter in this work is related to MTS in this case.
| 2133 |@word middle:3 seems:1 covariance:19 thereby:1 mention:1 tr:1 moment:1 reduction:1 contains:3 score:2 outperforms:1 comparing:2 dx:1 ws1:2 cruz:1 shape:1 analytic:2 plot:5 discrimination:1 generative:4 transposition:1 herbrich:1 attack:1 simpler:2 height:1 consists:1 prove:1 fitting:2 combine:1 manner:1 inter:1 expected:1 indeed:1 frequently:2 growing:1 relying:1 spherical:2 considering:1 becomes:1 begin:1 notation:1 what:3 developed:2 unobserved:1 transformation:2 pmed:1 every:1 ti:1 classifier:3 k2:2 uk:1 control:1 positive:7 before:1 local:1 modify:1 io:2 encoding:2 approximately:1 ap:1 might:1 pami:1 equivalence:1 co:1 range:2 acknowledgment:1 testing:1 practice:1 definite:5 oio:1 digit:2 significantly:1 confidence:1 radial:1 integrating:1 downsampled:1 cannot:3 unlabeled:12 selection:1 close:2 context:2 influence:3 applying:3 reviewer:1 straightforward:2 williams:2 l:2 sharpness:1 simplicity:2 immediately:1 continued:1 haussler:3 his:1 dw:1 ralf:1 embedding:2 notion:1 laplace:2 feel:1 us:1 hypothesis:3 element:1 storkey:1 labeled:9 database:1 observed:2 region:3 sonnenburg:1 prn:3 highest:1 mentioned:1 xio:5 ideally:1 dynamic:2 trained:1 depend:2 solving:1 deliver:1 upon:1 basis:1 vague:1 easily:1 joint:1 represented:3 talk:1 distinct:1 describe:3 london:1 approached:1 analyser:1 choosing:1 whose:1 richer:1 larger:1 say:1 yyy:2 neil:1 gp:3 itself:1 noisy:1 beal:2 rr:4 matthias:3 propose:3 product:5 remainder:1 combining:1 intuitive:1 los:1 exploiting:1 cluster:20 extending:1 produce:1 object:1 depending:1 ac:3 come:2 motoaki:1 tommi:2 closely:2 stochastic:1 wid:1 vx:1 material:1 implementing:1 require:1 preliminary:2 anonymous:1 normal:1 exp:1 lawrence:1 mapping:1 pointing:1 estimation:1 label:11 tool:1 weighted:2 amos:1 mit:3 clearly:1 gaussian:8 aim:2 rather:3 factorizes:1 jaakkola:2 encode:3 focus:1 yo:2 likelihood:1 indicates:2 seeger:5 contrast:1 baseline:4 sense:2 inference:2 flaw:1 typically:2 mediator:5 jq:1 interested:3 classification:11 flexible:1 colt:1 html:2 priori:4 smoothing:1 special:4 mutual:8 equal:1 once:2 construct:1 having:2 sampling:1 represents:1 future:1 report:6 nonsmooth:2 employ:2 familiar:1 ourselves:1 microsoft:1 tlx:1 highly:1 alignment:1 introduces:1 mixture:18 accurate:1 poorer:1 integral:2 closer:1 necessary:1 plugged:1 euclidean:2 taylor:1 abundant:1 walk:1 koji:1 tsuda:1 fitted:3 earlier:1 pia:1 introducing:2 entry:1 subset:1 successful:2 predicate:1 too:2 ukr:1 varies:1 density:5 probabilistic:2 pool:4 together:1 concrete:2 squared:1 again:2 zen:1 worse:1 toy:1 satisfy:1 depends:2 try:1 helped:1 responsibility:1 lot:1 start:1 bayes:1 slope:1 ass:1 il:1 generalize:1 bayesian:19 handwritten:2 published:1 ed:9 against:2 failure:1 thereof:1 mi:31 sampled:2 dataset:6 holdout:1 mitchell:1 knowledge:3 subsection:6 dimensionality:2 supervised:6 follow:1 tom:1 done:1 evaluated:1 strongly:2 furthermore:2 just:1 correlation:3 hand:1 trust:2 christopher:1 o:2 lack:1 believe:1 effect:2 matt:1 concept:2 true:3 normalized:1 former:1 analytically:2 inspiration:1 symmetric:1 deal:1 zaragoza:1 ll:1 hill:1 override:1 oid:1 performs:1 image:1 variational:11 recently:1 discriminants:3 behaves:2 mt:6 hugo:1 volume:1 jl:1 discussed:1 interpretation:2 extend:1 he:1 expressing:1 refer:1 ai:4 llx:2 tuning:1 approx:2 pm:6 analyzer:4 centre:2 gratefully:1 similarity:2 posterior:9 prime:1 scenario:4 disappointing:1 binary:2 success:1 came:1 muller:1 seen:7 dai:3 somewhat:3 purity:2 july:1 semi:3 smoother:2 afterwards:1 ii:6 smooth:1 technical:5 faster:1 cross:1 long:1 concerning:1 jy:1 ucsccrl:1 basic:2 essentially:2 metric:1 kernel:101 represent:1 interval:1 wealth:1 ratsch:1 source:2 rest:1 probably:1 suspect:1 tend:2 elegant:1 olx:2 incorporates:1 inconsistent:1 odds:2 extracting:1 ee:3 call:1 easy:2 rendering:1 fit:2 architecture:1 restrict:1 opposite:1 perfectly:1 inner:5 idea:4 tm:1 favour:2 motivated:1 expression:1 ltd:1 f:1 render:1 peter:1 hardly:1 remark:1 useful:2 santa:1 amount:1 nonparametric:1 transforms:1 inspiring:1 http:2 specifies:1 sl:1 discrete:3 gunnar:1 blum:1 drawn:1 tly:1 run:4 powerful:3 arrive:2 family:4 almost:1 forrest:1 separation:1 soeren:1 decision:1 scaling:4 vb:18 distinguish:1 fascinating:1 nonnegative:1 ri:1 flat:2 encodes:2 argument:2 martin:1 conjugate:2 across:1 describes:1 xlj:1 s1:1 hl:1 jeffreys:1 vq:1 mutually:1 discus:2 fail:1 mechanism:1 tractable:1 available:2 gaussians:2 epm:2 apply:4 kawanabe:1 generic:2 bii:1 anymore:1 robustness:1 original:1 inconsistently:1 denotes:2 clustering:1 include:1 ghahramani:1 build:1 bl:1 eh1:1 parametric:1 exclusive:1 diagonal:8 exhibit:1 distance:6 thank:1 chris:2 barber:1 discriminant:2 index:1 ql:1 unfortunately:1 rise:1 implementation:4 design:2 countable:1 unknown:1 perform:1 upper:3 convolution:1 datasets:1 markov:1 behave:1 ever:1 sharp:1 david:3 namely:1 pair:2 specified:2 required:1 california:1 learned:6 nip:3 trans:1 able:4 usually:1 xm:2 royal:1 natural:3 ne:1 acknowledges:1 mediated:1 prior:20 fully:1 generation:1 var:1 validation:1 degree:1 consistent:1 principle:1 plotting:1 occam:1 share:1 elsewhere:1 compatible:1 institute:1 sparse:1 edinburgh:2 boundary:1 overcome:1 curve:4 valid:2 ending:1 contour:3 world:1 studentship:1 author:3 adaptive:1 employing:1 approximate:4 global:1 correlating:1 assumed:1 discriminative:5 xi:2 latent:5 learn:3 correlated:1 klausrobert:1 du:11 expansion:1 diag:2 yianilos:1 sp:1 did:1 main:2 csd:1 motivation:1 noise:1 s2:4 eural:1 tl:1 definiteness:1 exponential:2 xl:3 lie:2 pe:1 watkins:2 cotraining:1 ix:2 embed:1 dl:4 workshop:1 mnist:3 avrim:1 nec:1 simply:2 likely:1 explore:1 partially:1 joined:1 dh:1 conditional:6 marked:1 rbf:10 towards:2 fisher:5 feasible:1 uniformly:1 called:3 partly:1 wio:1 holloway:1 support:3 latter:2 arises:1 szummer:1 princeton:1 ex:1 |
1,245 | 2,134 | Model-Free Least Squares Policy Iteration
Michail G. Lagoudakis
Department of Computer Science
Duke University
Durham, NC 27708
[email protected]
Ronald Parr
Department of Computer Science
Duke University
Durham, NC 27708
[email protected]
Abstract
We propose a new approach to reinforcement learning which combines
least squares function approximation with policy iteration. Our method
is model-free and completely off policy. We are motivated by the least
squares temporal difference learning algorithm (LSTD), which is known
for its efficient use of sample experiences compared to pure temporal
difference algorithms. LSTD is ideal for prediction problems, however it
heretofore has not had a straightforward application to control problems.
Moreover, approximations learned by LSTD are strongly influenced by
the visitation distribution over states. Our new algorithm, Least Squares
Policy Iteration (LSPI) addresses these issues. The result is an off-policy
method which can use (or reuse) data collected from any source. We have
tested LSPI on several problems, including a bicycle simulator in which
it learns to guide the bicycle to a goal efficiently by merely observing a
relatively small number of completely random trials.
1 Introduction
Linear least squares function approximators offer many advantages in the context of reinforcement learning. While their ability to generalize is less powerful than black box
methods such as neural networks, they have their virtues: They are easy to implement and
use, and their behavior is fairly transparent, both from an analysis standpoint and from
a debugging and feature engineering standpoint. When linear methods fail, it is usually
relatively easy to get some insight into why the failure has occurred.
Our enthusiasm for this approach is inspired by the least squares temporal difference learning algorithm (LSTD) [4]. LSTD makes efficient use of data and converges faster than
other conventional temporal difference learning methods. Although it is initially appealing
to attempt to use LSTD in the evaluation step of a policy iteration algorithm, this combination can be problematic. Koller and Parr [5] present an example where the combination of
LSTD style function approximation and policy iteration oscillates between two bad policies
in an MDP with just 4 states. This behavior is explained by the fact that linear approximation methods such as LSTD compute an approximation that is weighted by the state
visitation frequencies of the policy under evaluation. Further, even if this problem is overcome, a more serious difficulty is that the state value function that LSTD learns is of no use
for policy improvement when a model of the process is not available.
This paper introduces the Least Squares Policy Iteration (LSPI) algorithm, which extends
the benefits of LSTD to control problems. First, we introduce LSQ, an algorithm that
learns least squares approximations of the state-action ( ) value function, thus permitting
action selection and policy improvement without a model. Next we introduce LSPI which
uses the results of LSQ to form an approximate policy iteration algorithm. This algorithm
combines the policy search efficiency of policy iteration with the data efficiency of LSTD.
It is completely off policy and can, in principle, use data collected from any reasonable
sampling distribution. We have evaluated this method on several problems, including a
simulated bicycle control problem in which LSPI learns to guide the bicycle to the goal by
observing a relatively small number of completely random trials.
2 Markov Decision Processes
We assume that the underlying control problem is a Markov Decision Process (MDP). An
MDP is defined as a 4-tuple
where: is a finite set of states; is a finite set of
actions; is a Markovian transition model where
represents the probability of
going from state to state with action ; and is a reward function IR,
such that !
represents the reward obtained when taking action in state and
ending up in state " .
We will be assuming that the MDP has an infinite horizon and that future rewards are
discounted exponentially with a discount factor #%$'& ()+*"
. (If we assume that all policies
are proper, our results generalize to the undiscounted case.) A stationary policy , for an
MDP is a mapping ,-./43
10 , where ,2!"
is the action the agent takes at state . The
state-action value function
5
is defined over all possible combinations of states and
actions and indicates the expected, discounted total reward when taking action in state
and following policy , thereafter. The exact
-values for all state-action pairs can be
found by solving the linear system of the Bellman equations :
3
!6
7
28'9:!6
7
<;=#>@?BA !
3
! ,2
C
? A
M3
where 9:3R37
D8FE S3 GH
IJ!6
KL
. In matrix format, the
system becomes
8
3
9N;O#QP
9
, where
and
are vectors of size T DTUT 0MT and P
is a stochastic matrix of
3
describes the transitions from pairs 7
to pairs !@,2!H
@
.
size T DTLT 0 T7OT DTLT 0MT
. P
For every MDP, there exists an optimal policy, ,V , which maximizes the expected, discounted return of every state. Policy iteration is a method of discovering this policy by
iterating through a sequence of monotonically improving policies. Each iteration consists of two phases. Value determination computes the state-action values for a policy ,.WYX[Z by solving the above system. Policy improvement defines the next policy as
3fhgLi
,.WYX[\^]@ZC"
J8`_a
bc_d)e
7
. These steps are repeated until convergence to an optimal policy, often in a surprisingly small number of steps.
3 Least Squares Approximation of Q Functions
Policy iteration relies upon the solution of a system of linear equations to find the Q values
for the current policy. This43
is impractical for large state and action spaces. In such cases we
may wish to approximate
with a parametric function approximator and do some form
of approximate policy iteration. We now address the problem of finding a set of parameters
that maximizes the accuracy of our approximator. A common class of approximators is
the so called linear architectures, where the value function is approximated as a linear
weighted combination of
basis functions (features):
3
!
>
8
!6
7
8
!7
S
]
where is a set of weights (parameters). In general,
-T DTLT 0 T and so, the linear system
above now becomes an overconstrained system over the parameters :
# P
3
9F;O#QP
9
3
3
where is a
T DTLT 0MT!)
matrix. We are interested in a set3 of weights
that yields a fixed
3
point in value function space, that is a value function
that is invariant under
8
one step of value determination followed by orthogonal
projection to the space spanned by
the basis functions. Assuming that the columns of are linearly independent this is
]
3
9N;O#QP
3
3
8
8
3
8
3
# P
]
9
We note that this is the standard fixed point approximation method for linear value functions
with the exception3 that the problem is formulated in terms of Q values instead of state
values. For any P , the solution is guaranteed to exist for all but finitely many # [5].
4 LSQ: Learning the State-Action Value Function
3
In the previous section we assumed that a model M
P
of the underlying MDP is available. In many practical applications, such a model is not available and the value function
or, more precisely, its parameters have to be learned from sampled data. These sampled
data are tuples of the form:
L
, meaning that in state , action was taken, a reward was received, and the resulting state was . These data can be collected from actual
(sequential) episodes or from random queries to a generative model of the MDP. In the
extreme case, they can be experiences of other agents on the same MDP. We
know that
3
8! , where
the desired
set
of
weights
can
be
found
as
the
solution
of
the
system,
3
9 .
8
# P
and
8
3
The matrix P
and the vector 9 are unknown and so, and cannot be
a
3
determined
priori. However, and can be approximated using samples. Recall that , P
, and 9
are of the form:
FHG A:I
(*),+.-0/213-4 5
%&
&
&
%&
A
&
A
6 606
0
(*),+./7184 5
"$#
'
),+.-0/ 1J-0/+0KL42(?),+0KM/ N),+K4 4 5
@A
&
B
60606
(*),+:9 ;<9=/ 139 >?9=4 5
CED
G ATI
F
%&
&
'
&
'
F
B
@ A
A
A
606O6
),+:/71J/+ K 42UV),+./ 1J/ + K 4
B
606O6
),+T9 ;J9=/ 139 >?9W/+0KL42UV),+:9 ;<9X/ 139 >?9Y/+0K4
G A I
A
6O606
),+ 9 ;<9 /21 9 >*9 /+ K 42(?),+ K / NQ),+ K 4 4 5
),+ - /21 - /+ K 42UV),+ - /21 - /+ K 4
FHG A I
RS#
FPG A I
A
6O606
),+./21J/+ K 42(?),+ K / NQ),+ K 4 4 5
FPG A I
"$#
@A
Given a set of samples, Z 8\["^]0_GJ]0_
K]0_ :]0_@
+T<`M8 *6OaRbcb dfe , where the :]0_G
J]0_I
are sampled from according to distribution g and the 6] _ are sampled according to
3
K]0_ T :]0_GJ]0_I
, we can construct approximate versions of
,P
, and 9 as follows :
hi
i
8
i
j
bc
!:]m
<]m
i
o
p
P
q
3
8
i
j
nbo
!K]lk @,sr K
l
] k
t
i
o
bc
] _C
] _I
hi
nbo
]lk
]lk
!H]m @,
] _
t
r K
cb
r K
] m
t
i
o
o
cb
K] _ @,
hi
o
p
i
9
8
j
] k
l
bb
] _
bb
:]m
nbo
o
p
o
These approximations can be thought of as first sampling rows from
according to g
and then, conditioned on 3 these
samples,
as
sampling
terms
from
the
summations
in the
corresponding rows of P
and 9 . The sampling distribution from the summations is
governed by the underlying dynamics (
U
) of the process as the samples in Z are
taken directly from the MDP.
Given
,P
q
3
, and 9 ,
and can be approximated as
8
3
# P
q
and
9
8
With d uniformly
distributed samples over pairs of states and actions
mations and are consistent approximations of the true and :
d
8
T
TUT
T
and
The Markov property ensures that
the solution
3
sufficiently large d whenever
exists:
3
] C
8
8
d
T DTUT
T
3
!7
, the approxi-
d
28
T
TUT
T
will converge to the true solution
]
d
T DTUT
T
]
28
3
for
3
8$
In the more general case, where g is not uniform, we will compute a weighted projection,
which minimizes the g weighted distance in the projection step. Thus, state is implicitly
assigned weight g!K
and the projection minimizes the weighted sum of squared
errors
3
with respect to g . In LSTD, for example, g is the stationary distribution of P , giving high
weight to frequently visited states, and low weight to infrequently visited states.
As with LSTD, it is easy to see that approximations ( 6 ) derived from different
]
]
sets of samples (Z Z ) can be combined additively to yield a better approximation that
]
corresponds to the combined set of samples:
8
;
]
and
8
]
;
This observation
leads to an incremental update rule for and . Assume that initially 8
( and
8
( . For a fixed policy, a new sample !6
contributes to the approximation
according to the following update equation :
;
5
+
5
#
! ,2
@
and
.;
7
We call this new algorithm LSQ due to its similarity to LSTD. However, unlike LSTD, it
computes Q functions and does not expect the data to come from any particular Markov
chain. It is a feature of this algorithm that it can use the same set of samples to compute Q
values for any policy representation that offers an action choice for
each in the set. The
policy merely determines which !,2K!@,2!H
@
is added to for each sample. Thus,
LSQ can use every single sample available to it no matter what policy is under evaluation.
We note that if a particular set of projection
weights are desired, it is straightforward to
reweight the samples as they are added to .
Notice that apart from storing the samples, LSQ requires only J7
space independently
of the size of the state and the action
space. For each sample in Z , LSQ incurs a cost of
and and a one time cost of J7 "
to solve the system and
J
to update the matrices
find
the weights. Singular value decomposition (SVD) can be used for robust inversion of
as it is not always a full rank matrix.
LSQ includes LSTD as a special case where there is only one action available. It is also
possible to extend LSQ to LSQ( ) in a way that closely resembles LSTD( ) [3], but in
that case the sample set must consist of complete episodes generated using the policy under evaluation, which again raises the question of bias due to sampling distribution, and
prevents the reusability of samples. LSQ is also applicable in the case of infinite and continuous state and/or action spaces with no modification. States and actions are reflected
only through the basis functions of the linear approximation and the resulting value function can cover the entire state-action space with the appropriate set of continuous basis
functions.
5 LSPI: Least Squares Policy Iteration
The LSQ
algorithm provides a means of learning an approximate state-action value funcS3
tion,
!6
7
, for any fixed policy , . We now integrate LSQ into an approximate policy
iteration algorithm. Clearly, LSQ is a candidate for the value determination step. The key
insight is that we can achieve the policy improvement step without ever explicitly representing our policy and without any sort
of model. Recall that in policy improvement, ,WYX[\^]@Z
43
!7
. Since LSQ computes Q functions directly,
will pick the action that maximizes
we do not need a model to determine our improved policy; all the information we need is
contained implicitly in the weights parameterizing our Q functions 1:
, WUX[\^]IZ !
8
_a
bc_d
e
7
8
_a
b2c_d
e
!7
We close the loop simply by requiring that LSQ performs this maximization for each
when constructing the matrix for a policy. For very large or continuous action spaces,
explicit maximization over may be impractical. In such cases, some sort of global nonlinear optimization may be required to determine the optimal action.
Since LSPI uses LSQ to compute approximate Q functions, it can use any data source for
samples. A single set of samples may be used for the entire optimization, or additional samples may be acquired, either through trajectories or some other scheme, for each iteration
of policy iteration. We summarize the LSPI algorithm in Figure 1. As with any approximate policy iteration algorithm, the convergence of LSPI is not guaranteed. Approximate
policy iteration variants are typically analyzed in terms of a value function approximation
error and an action selection error [2]. LSPI does not require an approximate policy representation, e.g., a policy function or ?actor? architecture, removing one source of error.
Moreover, the direct computation of linear Q functions from any data source, including
stored data, allows the use of all available data to evaluate every policy, making the problem of minimizing value function approximation error more manageable.
6 Results
We initially tested LSPI on variants of the problematic MDP from Koller and Parr [5],
essentially simple chains of varying length. LSPI easily found the optimal policy within
a few iterations using actual trajectories. We also tested LSPI on the inverted pendulum
problem, where the task is to balance a pendulum in the upright position by moving the cart
to which it is attached. Using a simple set of basis functions and samples collected from
random episodes (starting in the upright position and following a purely random policy),
LSPI was able to find excellent policies using a few hundred such episodes [7].
Finally, we tried a bicycle balancing problem [12] in which the goal is to learn to balance
and ride a bicycle to a target position located 1 km away from the starting location. Initially,
the bicycle?s orientation is at an angle of 90 to the goal. The state description is a six
dimensional vector D
, where is the angle of the handlebar, is the vertical
1
This is the same principle that allows action selection without a model in Q-learning. To our
knowledge, this is the first application of this principle in an approximate policy iteration algorithm.
LSPI (
/( / //*N./
)
// ( : Number of basis functions
// : Basis functions
// : Discount factor
// N : Stopping criterion
N # N),+./ 4
(default:
// : Initial policy, given as ,
//
: Initial set of samples, possibly empty
#
N K #
N
repeat
Update
N # N K
/
/( / /*N
// Add/remove
samples, or leave unchanged
K
#
// K
/ /( /
/
// = LSQ ( K
)
// that is, ( )
)
)
until (
)
(optional)
N K
(
N=
PLSQ
N K
return
K #
// In essence,
#
N
// return
Figure 1: The LSPI algorithm.
angle of the bicycle, and is the angle of the
bicycle to the goal. The actions are the torque
applied to the handlebar (discretized to [ a7
(RC; a<e ) and the displacement of the rider
(discretized to [ (3 ( aR()G;(3 ( aJe ). In our experiments, actions are restricted to be either
or (or nothing) giving a total of 5 actions2 . The noise in the system is a uniformly
distributed term in & ( (8a7G;( (8a added to the displacement component of the action. The
dynamics of the bicycle are based on the model described by Randl?v and Alstr?m [12]
and the time step of the simulation is set to (3 (R* seconds.
The state-action value function !7
for a fixed action
combination of 20 basis functions:
<* D
D
is approximated by a linear
C
8 ,
8
where
for ( and
,
for ( . Note that the state variable is
completely ignored. This block of basis functions is repeated for each of the 5 actions, giving a total of 100 basis functions and weights. Training data were collected by initializing
the bicycle to a random state around the equilibrium position and running small episodes
of 20 steps each using a purely random policy. LSPI was applied on training sets of different sizes and the average performance is shown in Figure 2(a). We used the same data
set for each run of policy iteration and usually obtained convergence in 6 or 7 iterations.
Successful policies usually reached the goal in approximately 1 km total, near optimal performance. We also show an annotated set of trajectories to demonstrate the performance
improvement over multiple steps of policy iteration in Figure 2(b).
The following design decisions influenced the performance of LSPI on this problem: As is
typical with this problem, we used a shaping reward [10] for the distance to the goal. In this
case, we used (3 (R* of the net change (in meters) in the distance to the goal. We found that
when using full random trajectories, most of our sample points were not very useful; they
occurred after the bicycle had already entered into a ?death spiral? from which recovery
was impossible. This complicated our learning efforts by biasing the samples towards
hopeless parts of the space, so we decided to cut off trajectories after 20 steps. This created
an additional problem because there was no terminating reward signal to indicate failure.
We approximated this with an additional shaping reward, which was proportional to the
2
Results are similar for the full 9-action case, but required more training data.
100
200
6th iteration
(crash)
Starting
Position
90
80
0
Percentage of trials reaching the goal
3rd iteration
70
2nd iteration (crash)
Goal
60
?200
5th and 7th
iteration
50
40
4th and 8th
iteration
?400
30
20
?600
10
0
1st iteration
0
500
1000
1500
2000
Number of training episodes
2500
3000
?800
?200
0
200
400
600
800
1000
1200
(a)
(b)
Figure 2: The bicycle problem: (a) Percentage of final policies that reach the goal, averaged
over 200 runs of LSPI for each training set size; (b) A sample run of LSPI based on 2500
training trials. This run converged in 8 iterations. Note that iterations 5 and 7 had different
Q-values but very similar policies. This was true of iterations
4
and
8 as well. The weights
of the ninth differed from the eighth by less than *H( ] in
, indicate convergence.
The curves at the end of the trajectories indicating where the bicycle has looped back for a
second pass through the goal.
net change in the square of the vertical angle. This roughly approximated the likeliness
of falling at the end of a truncated trajectory. Finally, we used a discount of ( ( , which
seemed to yield more robust performance.
We admit to some slight unease about the amount of shaping and adjusting of parameters
that was required to obtain good results on this problem. To verify that we had not eliminated the learning problem entirely through shaping, we reran some experiments using a
discount of ( . In this case LSQ simply projects the immediate reward function into the
column space of the basis functions. If the problem were tweaked too much, acting to
maximize the projected immediate reward would be sufficient to obtain good performance.
On the contrary, these runs always produced immediate crashes in trials.
7 Discussion and Conclusions
We have presented a new, model-free approximate policy iteration algorithm called LSPI,
which is inspired by the LSTD algorithm. This algorithm is able to use either a stored
repository of samples or samples generated dynamically from trajectories. It performs
action selection and approximate policy iteration entirely in value function space, without
any need for model. In contrast to other approaches to approximate policy iteration, it does
not require any sort of approximate policy function.
In comparison to the memory based approach of Ormoneit and Sen [11], our method makes
stronger use of function approximation. Rather than using our samples to implicitly construct an approximate model using kernels, we operate entirely in value function space and
use our samples directly in the value function projection step. As noted by Boyan [3] the
matrix used by LSTD and LSPI can be viewed as an approximate, compressed model.
This is most compelling if the columns of are orthonormal. While this provides some
intuitions, a proper transition function cannot be reconstructed directly from , making a
possible interpretation of LSPI as a model based method an area for future research.
In comparison to direct policy search methods [9, 8, 1, 13, 6], we offer the strength of
policy iteration. Policy search methods typically make a large number of relatively small
steps of gradient-based policy updates to a parameterized policy function. Our use of policy
iteration generally results in a small number of very large steps directly in policy space.
Our experimental results demonstrate the potential of our method. We achieved good performance on the bicycle task using a very small number of randomly generated samples
that were reused across multiple steps of policy iteration. Achieving this level of performance with just a linear value function architecture did require some tweaking, but the
transparency of the linear architecture made the relevant issues much more salient than
would be the case with any ?black box? approach. We believe that the direct approach to
function approximation and data reuse taken by LSPI will make the algorithm an intuitive
and easy to use first choice for many reinforcement learning tasks. In future work, we plan
to investigate the application of our method to multi-agent systems and the use of density
estimation to control the projection weights in our function approximator.
Acknowledgments
We would like to thank J. Randl?v and P. Alstr?m for making their bicycle simulator available. We also thank C. Guestrin, D. Koller, U. Lerner and M. Littman for helpful discussions. The first author would like to thank the Lilian-Boudouri Foundation in Greece for
partial financial support.
References
[1] J. Baxter and P.Bartlett. Reinforcement learning in POMDP?s via direct gradient ascent. In
Proc. 17th International Conf. on Machine Learning, pages 41?48. Morgan Kaufmann, San
Francisco, CA, 2000.
[2] D. Bertsekas and J. Tsitsiklis. Neuro-Dynamic Programming. Athena Scientific, Belmont,
Massachusetts, 1996.
[3] Justin A. Boyan. Least-squares temporal difference learning. In I. Bratko and S. Dzeroski,
editors, Machine Learning: Proceedings of the Sixteenth International Conference, pages 49?
56. Morgan Kaufmann, San Francisco, CA, 1999.
[4] S. Bradtke and A. Barto. Linear least-squares algorithms for temporal difference learning.
Machine Learning, 22(1/2/3):33?57, 1996.
[5] D. Koller and R. Parr. Policy iteration for factored mdps. In Proceedings of the Sixteenth
Conference on Uncertainty in Artificial Intelligence (UAI-00). Morgan Kaufmann, 2000.
[6] V. Konda and J. Tsitsiklis. Actor-critic algorithms. In NIPS 2000 editors, editor, Advances in
Neural Information Processing Systems 12: Proceedings of the 1999 Conference. MIT Press,
2000.
[7] M. G. Lagoudakis and R. Parr. Model-Free Least-Squares policy iteration. Technical Report
CS-2001-05, Department of Computer Science, Duke University, December 2001.
[8] A. Ng and M. Jordan. PEGASUS: A policy search method for large MDPs and POMDPs.
In Proceedings of the Sixteenth Conference on Uncertainty in Artificial Intelligence (UAI-00).
Morgan Kaufmann, 2000.
[9] A. Ng, R. Parr, and D. Koller. Policy search via density estimation. In Advances in Neural
Information Processing Systems 12: Proceedings of the 1999 Conference. MIT Press, 2000.
[10] Andrew Y. Ng, Daishi Harada, and Stuart Russell. Policy invariance under reward transformations: theory and application to reward shaping. In Proc. 16th International Conf. on Machine
Learning, pages 278?287. Morgan Kaufmann, San Francisco, CA, 1999.
[11] D. Ormoneit and S. Sen. Kernel-based reinforcement learning. To appear, Machine Learning,
2001.
[12] J. Randl?v and P. Alstr?m. Learning to drive a bicycle using reinforcement learning and shaping. In The Fifteenth International Conference on Machine Learning, 1998. Morgan Kaufmann.
[13] R. Sutton, D. McAllester, S. Singh, and Y. Mansour. Policy gradient methods for reinforcement
learning with function approximation. In Advances in Neural Information Processing Systems
12: Proceedings of the 1999 Conference, 2000. MIT Press.
| 2134 |@word trial:5 repository:1 version:1 inversion:1 manageable:1 stronger:1 nd:1 reused:1 km:3 additively:1 heretofore:1 r:1 tried:1 decomposition:1 simulation:1 pick:1 incurs:1 initial:2 t7:1 ati:1 reran:1 current:1 must:1 ronald:1 belmont:1 remove:1 update:5 stationary:2 generative:1 discovering:1 intelligence:2 nq:2 mgl:1 provides:2 location:1 direct:4 consists:1 combine:2 introduce:2 acquired:1 expected:2 roughly:1 behavior:2 frequently:1 simulator:2 multi:1 discretized:2 bellman:1 m8:1 torque:1 inspired:2 discounted:3 alstr:3 actual:2 becomes:2 project:1 moreover:2 underlying:3 maximizes:3 tweaked:1 what:1 minimizes:2 finding:1 transformation:1 impractical:2 temporal:6 every:4 oscillates:1 control:5 appear:1 bertsekas:1 engineering:1 sutton:1 approximately:1 black:2 resembles:1 dynamically:1 averaged:1 decided:1 practical:1 acknowledgment:1 block:1 implement:1 handlebar:2 displacement:2 area:1 thought:1 projection:7 tweaking:1 get:1 cannot:2 close:1 selection:4 context:1 fpg:2 impossible:1 conventional:1 straightforward:2 starting:3 independently:1 pomdp:1 recovery:1 pure:1 factored:1 insight:2 rule:1 parameterizing:1 spanned:1 orthonormal:1 financial:1 likeliness:1 target:1 exact:1 duke:5 programming:1 us:2 infrequently:1 approximated:6 located:1 cut:1 initializing:1 ensures:1 episode:6 russell:1 intuition:1 reward:12 littman:1 dynamic:3 terminating:1 raise:1 solving:2 singh:1 purely:2 upon:1 efficiency:2 completely:5 basis:11 easily:1 query:1 artificial:2 solve:1 compressed:1 ability:1 final:1 advantage:1 sequence:1 net:2 sen:2 propose:1 relevant:1 loop:1 entered:1 achieve:1 sixteenth:3 description:1 intuitive:1 convergence:4 empty:1 undiscounted:1 incremental:1 converges:1 leave:1 andrew:1 dzeroski:1 finitely:1 received:1 c:3 come:1 indicate:2 closely:1 annotated:1 stochastic:1 mcallester:1 require:3 transparent:1 summation:2 sufficiently:1 around:1 equilibrium:1 bicycle:17 mapping:1 parr:7 estimation:2 proc:2 applicable:1 visited:2 weighted:5 mit:3 clearly:1 always:2 reaching:1 rather:1 mations:1 varying:1 barto:1 derived:1 improvement:6 rank:1 indicates:1 contrast:1 helpful:1 stopping:1 entire:2 typically:2 initially:4 koller:5 going:1 fhg:2 interested:1 issue:2 orientation:1 priori:1 plan:1 special:1 fairly:1 construct:2 ng:3 sampling:5 eliminated:1 represents:2 stuart:1 looped:1 future:3 report:1 serious:1 few:2 randomly:1 lerner:1 phase:1 attempt:1 investigate:1 evaluation:4 introduces:1 analyzed:1 extreme:1 tut:2 chain:2 tuple:1 partial:1 experience:2 orthogonal:1 desired:2 column:3 compelling:1 markovian:1 cover:1 ar:1 maximization:2 cost:2 uniform:1 hundred:1 harada:1 successful:1 too:1 stored:2 combined:2 st:1 density:2 international:4 off:4 squared:1 again:1 possibly:1 admit:1 conf:2 style:1 return:3 potential:1 includes:1 matter:1 explicitly:1 tion:1 observing:2 pendulum:2 reached:1 sort:3 complicated:1 square:14 ir:1 accuracy:1 kaufmann:6 efficiently:1 yield:3 generalize:2 produced:1 trajectory:8 pomdps:1 drive:1 converged:1 influenced:2 reach:1 whenever:1 failure:2 frequency:1 sampled:4 adjusting:1 massachusetts:1 recall:2 knowledge:1 overconstrained:1 shaping:6 greece:1 back:1 o6:2 reflected:1 improved:1 evaluated:1 box:2 strongly:1 just:2 until:2 nonlinear:1 a7:2 defines:1 scientific:1 believe:1 mdp:11 verify:1 true:3 requiring:1 assigned:1 death:1 essence:1 noted:1 criterion:1 complete:1 demonstrate:2 performs:2 bradtke:1 meaning:1 lagoudakis:2 common:1 mt:3 qp:3 enthusiasm:1 attached:1 exponentially:1 extend:1 occurred:2 interpretation:1 slight:1 rd:1 uv:2 had:4 moving:1 ride:1 actor:2 similarity:1 add:1 apart:1 approximators:2 inverted:1 ced:1 guestrin:1 additional:3 morgan:6 michail:1 converge:1 determine:2 maximize:1 monotonically:1 signal:1 full:3 multiple:2 transparency:1 technical:1 faster:1 determination:3 b2c:1 offer:3 permitting:1 prediction:1 variant:2 neuro:1 essentially:1 fifteenth:1 iteration:40 kernel:2 achieved:1 set3:1 crash:3 randl:3 singular:1 source:4 standpoint:2 ot:1 operate:1 unlike:1 sr:1 ascent:1 cart:1 december:1 contrary:1 jordan:1 call:1 near:1 ideal:1 easy:4 spiral:1 baxter:1 architecture:4 motivated:1 six:1 bartlett:1 reuse:2 effort:1 action:35 ignored:1 useful:1 iterating:1 generally:1 j9:1 amount:1 discount:4 exist:1 percentage:2 problematic:2 s3:1 notice:1 iz:1 visitation:2 thereafter:1 key:1 salient:1 falling:1 achieving:1 merely:2 sum:1 run:5 angle:5 parameterized:1 powerful:1 uncertainty:2 extends:1 reasonable:1 decision:3 entirely:3 hi:3 daishi:1 followed:1 guaranteed:2 strength:1 precisely:1 relatively:4 format:1 department:3 according:4 debugging:1 combination:5 describes:1 rider:1 across:1 appealing:1 modification:1 making:3 explained:1 invariant:1 restricted:1 taken:3 equation:3 fail:1 know:1 end:2 available:7 away:1 appropriate:1 running:1 dfe:1 konda:1 giving:3 lspi:24 unchanged:1 added:3 question:1 already:1 pegasus:1 parametric:1 gradient:3 distance:3 thank:3 simulated:1 athena:1 collected:5 assuming:2 length:1 minimizing:1 balance:2 nc:2 reweight:1 ba:1 design:1 proper:2 policy:78 unknown:1 vertical:2 observation:1 markov:4 finite:2 optional:1 truncated:1 immediate:3 ever:1 mansour:1 ninth:1 wyx:3 pair:4 required:3 learned:2 nip:1 address:2 able:2 justin:1 usually:3 eighth:1 biasing:1 summarize:1 including:3 memory:1 difficulty:1 boyan:2 ormoneit:2 bratko:1 representing:1 scheme:1 mdps:2 lk:3 created:1 meter:1 expect:1 j8:1 proportional:1 approximator:3 foundation:1 integrate:1 agent:3 sufficient:1 consistent:1 principle:3 editor:3 storing:1 critic:1 balancing:1 row:2 hopeless:1 surprisingly:1 repeat:1 free:4 zc:1 guide:2 bias:1 tsitsiklis:2 taking:2 benefit:1 distributed:2 overcome:1 default:1 curve:1 transition:3 ending:1 computes:3 seemed:1 author:1 made:1 reinforcement:7 projected:1 san:3 reconstructed:1 approximate:17 implicitly:3 approxi:1 global:1 uai:2 assumed:1 francisco:3 tuples:1 search:5 continuous:3 why:1 lsq:19 learn:1 robust:2 ca:3 contributes:1 improving:1 excellent:1 constructing:1 did:1 linearly:1 noise:1 nothing:1 repeated:2 differed:1 position:5 wish:1 explicit:1 candidate:1 governed:1 learns:4 removing:1 bad:1 virtue:1 t9:1 exists:2 consist:1 sequential:1 conditioned:1 horizon:1 durham:2 simply:2 prevents:1 contained:1 lstd:19 corresponds:1 determines:1 relies:1 goal:12 formulated:1 viewed:1 towards:1 change:2 infinite:2 determined:1 uniformly:2 upright:2 typical:1 acting:1 called:2 total:4 pas:1 invariance:1 svd:1 experimental:1 m3:1 indicating:1 support:1 evaluate:1 tested:3 |
1,246 | 2,135 | Novel iteration schemes for the Cluster
Variation Method
Hilbert J. Kappen
Department of Biophysics
Nijmegen University
Nijmegen, the Netherlands
bert?mbfys.kun.nl
Wim Wiegerinck
Department of Biophysics
Nijmegen University
Nijmegen, the Netherlands
wimw?mbfys.kun.nl
Abstract
The Cluster Variation method is a class of approximation methods containing the Bethe and Kikuchi approximations as special
cases. We derive two novel iteration schemes for the Cluster Variation Method. One is a fixed point iteration scheme which gives a
significant improvement over loopy BP, mean field and TAP methods on directed graphical models. The other is a gradient based
method, that is guaranteed to converge and is shown to give useful
results on random graphs with mild frustration. We conclude that
the methods are of significant practical value for large inference
problems.
1
Introduction
Belief Propagation (BP) is a message passing scheme, which is known to yield exact
inference in tree structured graphical models [1]. It has been noted by several
authors that Belief Propagation can can also give impressive results for graphs that
are not trees [2].
The Cluster Variation Method (CVM), is a method that has been developed in the
physics community for approximate inference in the Ising model [3]. The CVM approximates the joint probability distribution by a number of (overlapping) marginal
distributions (clusters). The quality of the approximation is determined by the size
and number of clusters. When the clusters consist of only two variables, the method
is known as the Bethe approximation. Recently, the method has been introduced
by Yedidia et a1.[4] into the machine learning community, showing that in the Bethe
approximation, the CVM solution coincides with the fixed points of the belief propagation algorithm. For clusters consisting of more than two variables, [4] present a
message passing scheme called generalized belief propagation (GBP). This approximation to the free energy is often referred to as the Kikuchi approximation. They
show, that GBP gives a significant improvement over the Bethe approximation for
a small two dimensional Ising lattice with random couplings. However, for larger
latices, both GBP and BP fail to converge [4, 5].
In [5] the CCCP method is proposed, which is a double loop iteration algorithm that
is guaranteed to converge for the general CVM problem. Intuitively, the method
consists of iteration a sequence of convex subproblem (outer loop) each of which is
solved using a fixed point iteration method (inner loop). In this sense, the method
is similar to the UPS algorithm of [6] which identifies trees as subproblems.
In this paper, we propose two algorithms, one is a fixed point iteration procedure, the other a gradient based method. We show that the fixed point iteration
method gives very fast convergence and accurate results for some classical directed
graphical models. However, for more challenging cases the fixed point method does
not converge and the gradient based approach, which is guaranteed to converge, is
preferable.
2
The Cluster Variation Method
In this section, we briefly present the cluster variation method. For a more complete
treatment see for instance [7]. Let x = (Xl, ... ,xn ) be a set of variables, where each
Xi can take a finite number of values. Consider a probability distribution on X of
the form
1_ -H(x)
( ) __
Z =
e-H(x)
PH X - Z(H)e
2:=
x
It is well known, that PH can be obtained as the minimum of the free energy, which
is a functional over probability distributions of the following form:
FH(P)
= (H) + (logp) ,
(1)
where the expectation value is taken with respect to the distribution p , i.e. (H) =
L x P(x)H(x). When one minimizes FH(P) with respect to P under the constraint
of normalization L x P(X) = 1, one obtains PH.
Computing marginals of PH such as PH(Xi) or PH(Xi, Xj) involves sums over all
states, which is intractable for large n. Therefore, one needs tractable approximations to PH. The cluster variation method replaces the probability distribution
PH(X) by a large number of (possibly overlapping) probability distributions , each
describing a sub set (cluster) of variables. Due to the one-to-one correspondence
between a probability distribution and the minima of a free energy we can define approximate probability distributions by constructing approximate free energies and
computing their minimum. This is achieved by approximating Eq. 1 in terms of the
cluster probabilities. The solution is obtained by minimizing this approximate free
energy subject to normalization and consistency constraints.
Define clusters as subsets of distinct variables: Xa = (XiI' ... ,Xik), with 1 ~ i j ~ n.
Consider the set of clusters P that describe the interactions in H and write H as a
sum of these interactions:
H(x) =
Hl(xoJ
2:=
a EP
We now define a set of clusters B, that will determine our approximation in the
cluster variation method. For each cluster a E B, we introduce a probability
distribution Pa(xa) which jointly must approximate p(x). B should at least contain
the interactions in p(x) in the following way: Va E P => 30:' E B,a c a'. In
addition, we demand that no two clusters in B contain each other: a, a' E B =>
a rt a', a' rt a. The minimal choice for B is to chose clusters from P itself. The
maximal choice for B is the cliques obtained when constructing the junction tree[8].
In this case, the clusters in B form a tree structure and the CVM method is exact.
Define a set of clusters M that consist of any intersection of a number of clusters
of B: M = {,BI,B = nkak, ak E B}, and define U = BuM. Once U is given, we
define numbers a/3 recursively by the Moebius formula
1=
L
ao;,
o;EU,o;"J/3
In particular, this shows that ao; = 1 for
0: E
V(3 E U
B.
The Moebius formula allows us to rewrite (H) in terms of the cluster probabilities
(H) = Lao; LPo;(xo;)Ho;(xo;),
o;EU
x"
(2)
with Ho;(xo;) = L./3EP,/3co; Hh(X/3) . Since interactions Hh may appear in more than
one Ho;, the constants ao; ensure that double counting is compensated for.
Whereas (H) can be written exactly in terms of Po;, this is not the case for the
entropy term in Eq. 1. The approach is to decompose the entropy of a cluster 0: in
terms of 'connected entropies' in the following way: 1
(3)
x"
/3Co;
where the sum over (3 contains all sub clusters of 0:. Such a decomposition can be
made for any cluster. In particular it can be made for the 'cluster' consisting of all
variables, so that we obtain
S = - LP(x) logp(x) = L Sh?
x
/3
(4)
The cluster variation method approximates the total entropy by restricting this
latter sum to only clusters in U and re-expressing Sh in terms of So;, using the
Moebius formula and the definition Eq. 3.
(5)
/3EU
/3EU 0;"J/3
o;EU
Since So; is a function of Po; (Eq. 3) , we have expressed the entropy in terms of
cluster probabilities Po; .
The quality of this approximation is illustrated in Fig. 1 for the SK model. Note,
that both the Bethe and Kikuchi approximation strongly deteriorate around J = 1,
which is where the spin-glass phase starts. For J < 1, the Kikuchi approximation is
superior to the Bethe approximation. Note, however, that this figure only illustrates
the quality of the truncations in Eq. 5 assuming that the exact marginals are known.
It does not say anything about the accuracy of the approximate marginals using
the approximate free energy.
Substituting Eqs. 2 and 5 into the free energy Eq. 1 we obtain the approximate
free energy of the Cluster Variation method. This free energy must be minimized
subject to normalization constraints L.x" Po; (x o; ) = 1 and consistency constraints
Po;(X/3) = P/3(X/3),
with Po; (X/3) = L. x
"\f3
0:,(3 E U,(3 C 0:.
(6)
Po; (xo;).
IThis decomposition is similar to writing a correlation in terms of means and covariance.
For instance when a = (i) , S(i) = SIi) is the usual mean field entropy and S(ij) = Sli) +
SIj)
+ Slij)
defines the two node correction Slij)"
12
10
8
>a.
e
c
6
lJ.J
4
"-
2
"-
"-
"-
0
0.5
1.5
2
J
Figure 1: Exact and approximate entropies for the fully connected Boltzmann-Gibbs
distribution on n = 10 variables with random couplings (SK model) as a function of mean
coupling strength. Couplings Wij are chosen from a normal Gaussian distribution with
mean zero and standard deviation J /..;n. External fields ()i are chosen from a normal
Gaussian distribution with mean zero and standard deviation 0.1. The exact entropy is
computed from Eq. 4. The Bethe and Kikuchi approximations are computed using the
approximate entropy expression Eq. 5 with exact marginals and by choosing B as the set
of all pairs and all triplets, respectively.
The set of consistency constraints can be significantly reduced because some constraints imply others. Let 0:,0:', .. . denote clusters in Band fJ, fJ', ... denote clusters
in M.
? If fJ
c fJ' Co: and Pa(x/3') = P/3' (x/3') and Pa(x/3 ) = P/3(x/3), then P/3' (x/3) =
P/3 (X/3)' This means that constraints between clusters in M can be removed .
fJ c fJ' c 0: , 0:' and Pa(x/3') = Pa' (x/3') and p,,,(x/3) = P/3 (x /3 ), then
Pa,(x/3) = P/3 (x/3)' This means that some constraints between clusters in B
? If
and M can be removed.
We denote the remaining necessary constraints by
0:
---t
fJ.
Adding Lagrange multipliers for the constraints we obtain the Cluster Variation
free energy:
aEU
- L Aa (LPa(Xa) aEU
x"
1) -
L
/3 EM
x"
L L Aa/3 (X/3) (Pa(x /3 ) - P/3 (x/3))
a-+/3 X f3
(7)
3
Iterating Lagrange multipliers
c(vm), a E U equal to zero, one can express the cluster probabilities in
By setting 88F
PO! X o:
terms of the Lagrange multipliers:
exp (-Ha(Xa)
;
+ L )..a(3 (X(3))
a
exp (-H(3 (X(3 ) - a1 L
;
(8)
(3 f-a
(3
)..a(3 (X(3 ))
(9)
(3 a-t (3
The remaining task is to solve for the Lagrange multipliers such that all constraints
(Eq. 6) are satisfied. We present two ways to do this.
When one substitutes Eqs. 8-9 into the constraint Eqs. 6 one obtains a system of
coupled non-linear equations. In Yedidia et al.[4] a message passing algorithm was
proposed to find a solution to this problem. Here, we will present an alternative
method, that solves directly in terms of the Lagrange multipliers.
3.1
Fixed point iteration
Consider the constraints Eq. 6 for some fixed cluster fJ and all clusters a -+ fJ and
define B(3 = {a E Bla -+ fJ }? We wish to solve for all constraints a -+ fJ, with
a E B(3 by adjusting )..a(3, a E B(3. This is a sub-problem with IB(3 IIX(3 I equations
and an equal number of unknowns, where IB(3 1 is the number of elements of B(3
and IX(3 1 is the number of values that x(3 can take. The probability distribution P(3
(Eq. 9) depends only on these Lagrange multipliers. Pa (Eq. 8) depends also on other
Lagrange multipliers. However, we consider only its dependence on )..a(3 , a E B(3 ,
and consider all other Lagrange multipliers as fixed . Thus,
(10)
with Pa independent of ).. a(3, a E B(3 .
Substituting, Eqs. 9 and 10 into Eq. 6, we obtain a set of linear equations for
)..a(3 (x(3 ) which we can solve in closed form:
)..a(3 (X(3 ) = -
a(3
alB IH(3 (X(3 ) - L A aa do gPa (X(3 )
l
+
(3
a'
with
Aaa =
l
1
/ja a l
-
--c=--:-
a(3
+ IB(3 1
We update the probabilities with the new values of the Lagrange multipliers using
Eqs. 9 and 10. We repeat the above procedure for all fJ E M until convergence.
3.2
Gradient descent
We define an auxiliary cost function
C = L LP(3 (X(3 ) log P(3 ((X(3 )) = L Ca(3
a(3 Xf3
Pa x(3
a(3
(11)
that is zero when all constraints are satisfied and positive otherwise and minimize
this cost function with respect to the Lagrange multipliers )..a(3 (X(3 ). The gradient
of C is given by:
8C
Pf3(Xf3
-Pf3(Xf3
- -) ""'
~ (log
( )) - Calf3 ) - ""'
~ (PIaf3' (
xf3) - Pa ())
xf3
af3
a/-tf3
Pa' Xf3
13 ' +--a
with
4
4.1
Numerical results
Directed Graphical models
We show the performance of the fixed point iteration procedure on several 'real
world' directed graphical models. In figure 2a, we plot the exact single node
marginals against the approximate marginals for the Asia problem [8]. Clusters
in B are defined according to the conditional probability tables. Convergence was
reached in 6 iterations using fixed point iteration. Maximal error on the marginals
is 0.0033. For comparison, we computed the mean field and TAP approximations,
as previously introduced by [9]. Although TAP is significantly better than MF, it
is far worse than the CVM method. This is not surprising, since both the MF and
TAP approximation are based on single node approximation, whereas the CVM
method uses potentials up to size 3.
In figure 2b, we plot the exact single node marginals against the approximate CVM
marginals for the alarm network [10]. The structure and CPTs were downloaded
from www.cs.huji.ac.il;-nir. Clusters in B are defined according to the conditional probability tables and maximally contain 5 variables. Convergence was
reached in 15 iterations using fixed point iteration. Maximal error on the marginals
is 0.029. Ordinary loopy BP gives an error in the marginals of approximately 0.25
[2]. Mean field and TAP methods did not give reproducible results on this problem.
0.5 , - - - - - - - - - ---.--f'l---+
..' .'
(fj
0.4
0.8
coc
(fj
co
.a, 0.6
?~0 . 3
C1l
x
x
E
x
ec. 0.2
0.1
E
:2 0.4
>
c.
<t:
Co
()
+
0.2
-F?
.'
.'
,...
OL-----------~
o
Exact marginals
(a) Asia problem (n
= 8).
0.5
Exact marginals
(b) Alarm problem (n
= 37).
Figure 2: Comparison of single node marginals on two real world problems.
Finally, we tested the cluster variation method on randomly generated directed
graphical models. Each node is randomly connected to k parents. The entries of
the probability tables are randomly generated between zero and one. Due to the
large number of loops in the graph, the exact method requires exponential time in
the maximum clique size, which can be seen from Table 1 to scale approximately
linear with the network size. Therefore exact computation is only feasible for small
graphs (up to size n = 40 in this case).
For the CVM, clusters in B are defined according to the conditional probability
tables. Therefore, maximal cluster size is k + 1. On these more challenging cases,
the fixed point iteration method does not converge. The results shown are obtained
with conjugate gradient descent on the auxiliary cost function Eq. 11. The results
are shown in Table 1.
n
10
20
30
40
50
Iter
16
189
157
148
132
IGI
8
12
16
21
26
Potential error
0.018
0.019
0.033
0.048
-
Margin error
0.004
0.029
0.130
0.144
-
G
9.7e-ll
2.4e-4
2.1e-3
3.6e-3
4.5e-3
Table 1: Comparison of CYM method for large directed graphical models. Each node
is connected to k = 5 parents. IGI is the tree width of the triangulated graph required
for the exact computation. Iter is the number of conjugate gradient descent iterations of
the CYM method. Potential error and margin error are the maximum absolute distance
(MAD) in any of the cluster probabilities and single variable marginals computed with
CYM, respectively. G is given by Eq. 11 after termination of CYM.
4.2
Markov networks
We compare the Bethe and Kikuchi approximations for the SK model with n = 5
neurons as defined in Fig. 1. We expect that for small J the CVM approximation
gives accurate results and deteriorates for larger J.
We compare the Bethe approximation, where we define clusters for all pairs of nodes
and a Kikuchi approximation where we define clusters for all sub sets of three nodes.
The results are given in Table 2. We see that for the Bethe approximation, the
results of the fixed point iteration method (FPI) and the gradient based approach
agree. For the Kikuchi approximation the fixed point iteration method does not
converge and results are omitted. As expected, the Kikuchi approximation gives
more accurate results than the Bethe approximation for small J.
5
Conclusion
We have presented two iteration schemes for finding the minimum of the constraint
problem Eq. 7. One method is a fixed point iteration method that is equivalent
to belief propagation for pairwise interactions. This method is very fast and gives
very accurate results for 'not too complex' graphical models , such as real world
directed graphical models and frustrated Boltzmann distributions in the Bethe approximation. However, for more complex graphs such as random directed graphs or
more complex approximations, such as the Kikuchi approximation, the fixed point
iteration method does not converge. Empirically, it is found that smoothing may
somewhat help , but certainly does not solve this problem. For these more complex problems we propose to minimize an auxiliary cost function using a gradient
Bethe
FPI
J
0.25
0.50
0.75
1.00
1.50
2.00
Iter
Error
7
9
13
17
38
75
0.000161
0.001297
0.004325
0.009765
0.027217
0.049955
gradient
Iter
Error
7
11
14
15
16
20
0.000548
0.001263
0.004392
0.009827
0.027323
0.049831
Kikuchi
gradient
Iter
Error
120
221
86
49
150
137
0.000012
0.000355
0.021176
0.036882
0.059977
0.088481
Table 2: Comparison of Bethe and Kikuchi approximation for Boltzmann distributions.
Iter is the number of iterations needed. Error is the MAD of single variable marginals.
based method. Clearly, this approach is guaranteed to converge. Empirically, we
have found no problems with local minima. However , we have found that obtaining
solut ion with C close to zero may require many iterations.
Acknowledgments
This research was supported in part by the Dutch Technology Foundation (STW).
I would like to thank Taylan Cemgil for providing his Matlab graphical models
toolkit and Sebino Stramaglia (Bari, Italy) for useful discussions.
References
[1] J. Pearl. Probabilistic reasoning in intelligent systems: Networks of Plausible Inference. Morgan Kaufmann, San Francisco, California, 1988.
[2] Kevin P. Murphy, Yair Weiss, and Michael I. Jordan . Loopy belief propagation for
approximate inference: An empirical study. In Proceedings of Uncertainty in AI,
pages 467- 475, 1999.
[3] R. Kikuchi. Physical R eview, 81:988, 1951.
[4] J.S. Yedidia, W.T. Freeman, and Y. Weiss. Generalized belief propagation. In T.K.
Leen , T.G. Dietterich, and V. Tresp, editors, Advances in Neural Information Processing Systems 13 (Proceedings of th e 2000 Conference), 2001. In press.
[5] A.L. Yuille and A. Rangarajan. The convex-concave principle. In T.G. Dieterich,
S. Becker, and Z. Ghahramani, editors, Advances in Neural Information Processing
Systems, volume 14, 2002. In press.
[6] Y. Teh and M. Welling. The unified propagation and scaling algorithm. In T.G.
Dieterich, S. Becker, and Z. Ghahramani, editors, Advances in Neural Information
Processing Systems, volume 14, 2002. In press.
[7] H.J. Kappen. The cluster variation method for approximate reasoning in medical
diagnosis. In G. Nardulli and S. Stramaglia, editors, Modeling Bio-medical signals.
World-Scientific, 2002. In press.
[8] S.L. Lauritzen and D.J. Spiegelhalter. Local computations with probabilties on graphical structures and their application to expert systems. J. Royal Statistical society B ,
50:154- 227, 1988.
[9] H.J . Kappen and W .A.J.J. Wiegerinck. Second order approximations for probability
models. In Todd Leen, Tom Dietterich, Rich Caruana, and Virginia de Sa, editors,
Advances in Neural Information Processing Systems 13, pages 238- 244. MIT Press,
2001.
[10] 1. Beinlich, G. Suermondt, R. Chaves, and G. Cooper. The alarm monitoring system:
A case study with two probabilistic inference techniques for belief networks. In 2'nd
European Conference on AI in Medicin e, 1989.
| 2135 |@word mild:1 briefly:1 nd:1 termination:1 decomposition:2 covariance:1 recursively:1 kappen:3 contains:1 surprising:1 must:2 written:1 suermondt:1 numerical:1 plot:2 reproducible:1 update:1 node:9 af3:1 sii:1 consists:1 introduce:1 pairwise:1 deteriorate:1 expected:1 mbfys:2 ol:1 freeman:1 minimizes:1 developed:1 unified:1 finding:1 concave:1 preferable:1 exactly:1 bio:1 medical:2 appear:1 positive:1 local:2 todd:1 cemgil:1 ak:1 approximately:2 chose:1 challenging:2 co:5 bi:1 directed:8 practical:1 acknowledgment:1 bla:1 procedure:3 empirical:1 significantly:2 ups:1 close:1 writing:1 www:1 equivalent:1 compensated:1 convex:2 his:1 variation:13 exact:13 us:1 pa:12 element:1 bari:1 ising:2 ep:2 subproblem:1 solved:1 connected:4 eu:5 removed:2 stramaglia:2 rewrite:1 c1l:1 yuille:1 po:8 joint:1 distinct:1 fast:2 describe:1 kevin:1 choosing:1 larger:2 solve:4 plausible:1 say:1 otherwise:1 dieterich:2 jointly:1 itself:1 sequence:1 propose:2 interaction:5 maximal:4 loop:4 convergence:4 cluster:49 double:2 parent:2 rangarajan:1 kikuchi:13 help:1 derive:1 coupling:4 ac:1 ij:1 lauritzen:1 sa:1 eq:21 solves:1 auxiliary:3 c:1 involves:1 triangulated:1 ja:1 require:1 ao:3 decompose:1 correction:1 around:1 normal:2 exp:2 taylan:1 substituting:2 omitted:1 fh:2 wim:1 mit:1 clearly:1 gaussian:2 improvement:2 sense:1 glass:1 inference:6 lj:1 wij:1 smoothing:1 special:1 marginal:1 field:5 once:1 f3:2 equal:2 minimized:1 others:1 intelligent:1 randomly:3 murphy:1 phase:1 consisting:2 message:3 certainly:1 sh:2 nl:2 gpa:1 bum:1 aeu:2 accurate:4 necessary:1 tree:6 re:1 tf3:1 minimal:1 instance:2 modeling:1 caruana:1 logp:2 lattice:1 loopy:3 cost:4 ordinary:1 deviation:2 subset:1 entry:1 too:1 virginia:1 coc:1 huji:1 probabilistic:2 physic:1 vm:1 michael:1 frustration:1 satisfied:2 containing:1 possibly:1 worse:1 external:1 expert:1 potential:3 de:1 wimw:1 igi:2 depends:2 cpts:1 closed:1 xf3:6 reached:2 start:1 minimize:2 il:1 spin:1 accuracy:1 kaufmann:1 yield:1 monitoring:1 xoj:1 definition:1 against:2 energy:10 latex:1 treatment:1 adjusting:1 hilbert:1 eview:1 asia:2 tom:1 maximally:1 wei:2 leen:2 strongly:1 xa:4 correlation:1 until:1 overlapping:2 propagation:8 defines:1 quality:3 alb:1 scientific:1 dietterich:2 contain:3 multiplier:10 illustrated:1 ll:1 width:1 noted:1 anything:1 coincides:1 generalized:2 complete:1 fj:14 reasoning:2 novel:2 recently:1 superior:1 functional:1 empirically:2 physical:1 volume:2 approximates:2 marginals:16 significant:3 expressing:1 gibbs:1 ai:2 consistency:3 toolkit:1 impressive:1 italy:1 lpo:1 seen:1 minimum:5 morgan:1 somewhat:1 converge:9 determine:1 signal:1 cccp:1 a1:2 biophysics:2 va:1 expectation:1 dutch:1 iteration:23 normalization:3 achieved:1 ion:1 whereas:2 addition:1 subject:2 jordan:1 counting:1 xj:1 inner:1 expression:1 becker:2 passing:3 matlab:1 useful:2 iterating:1 netherlands:2 band:1 ph:8 reduced:1 deteriorates:1 diagnosis:1 xii:1 write:1 express:1 iter:6 graph:7 sum:4 uncertainty:1 fpi:2 scaling:1 guaranteed:4 correspondence:1 replaces:1 strength:1 lpa:1 constraint:16 bp:4 slij:2 department:2 structured:1 according:3 conjugate:2 em:1 lp:2 aaa:1 hl:1 intuitively:1 sij:1 xo:4 taken:1 equation:3 agree:1 previously:1 describing:1 fail:1 hh:2 needed:1 tractable:1 junction:1 yedidia:3 alternative:1 yair:1 ho:3 substitute:1 remaining:2 ensure:1 graphical:11 iix:1 moebius:3 ghahramani:2 approximating:1 classical:1 society:1 rt:2 usual:1 dependence:1 gradient:11 distance:1 thank:1 outer:1 mad:2 assuming:1 providing:1 minimizing:1 kun:2 subproblems:1 xik:1 nijmegen:4 stw:1 boltzmann:3 unknown:1 teh:1 neuron:1 markov:1 beinlich:1 finite:1 descent:3 bert:1 community:2 introduced:2 pair:2 required:1 gbp:3 tap:5 california:1 pearl:1 solut:1 royal:1 belief:8 scheme:6 technology:1 lao:1 imply:1 spiegelhalter:1 identifies:1 coupled:1 tresp:1 nir:1 fully:1 expect:1 foundation:1 downloaded:1 principle:1 editor:5 repeat:1 supported:1 free:10 truncation:1 absolute:1 xn:1 world:4 rich:1 author:1 made:2 san:1 sli:1 far:1 ec:1 welling:1 approximate:14 obtains:2 clique:2 conclude:1 francisco:1 xi:3 triplet:1 sk:3 table:9 bethe:14 ca:1 obtaining:1 complex:4 european:1 constructing:2 did:1 alarm:3 fig:2 referred:1 cvm:10 cooper:1 sub:4 wish:1 exponential:1 xl:1 ib:3 ix:1 formula:3 showing:1 consist:2 intractable:1 ih:1 restricting:1 adding:1 illustrates:1 demand:1 margin:2 mf:2 entropy:9 intersection:1 lagrange:10 expressed:1 aa:3 frustrated:1 conditional:3 feasible:1 determined:1 ithis:1 wiegerinck:2 called:1 total:1 latter:1 probabilties:1 tested:1 |
1,247 | 2,136 | Generalization Performance of Some Learning
Problems in Hilbert Functional Spaces
Tong Zhang
IBM T.J. Watson Research Center
Yorktown Heights, NY 10598
[email protected]
Abstract
We investigate the generalization performance of some learning problems in Hilbert functional Spaces. We introduce a notion of convergence
of the estimated functional predictor to the best underlying predictor, and
obtain an estimate on the rate of the convergence. This estimate allows
us to derive generalization bounds on some learning formulations.
1 Introduction
!"$#&%' (
)
In order to estimate a good predictor from a set of training data
randomly drawn
from , it is necessary to start with a model of the functional relationship. In this paper,
we consider models that are subsets in some Hilbert functional spaces * . Denote by + +-,
the norm in * , we consider models in the set ./10 324*657+ 8+ ,:9;=< , where
; is a parameter that can be used to control the size of the underlying model family. We
would like to find the best model in . which is given by:
> ?@ ACB7DFIKJMEHL G NH !8"O
()
(1)
By introducing a non-negative Lagrangian multiplier PQSR , we may rewrite the above
problem as:
> T@ ACB7IKDFJ E,G U NH !8"V
(
W PX + Y+Z,\[ )
(2)
We shall only consider this equivalent formulation in this paper. In addition, for technical
reasons, we also assume that ^]
C_ is a convex function of ] .
In machine learning, our goal is often to predict an unobserved output value based on an
observed input vector . This requires us to estimate a functional relationship
from a set of example pairs of
. Usually the quality of the predictor
can be
that is problem dependent. In machine learning,
measured by a loss function
we assume that the data
are drawn from an underlying distribution which is not
so that the expected true loss of given below is as small
known. Our goal is to find
as possible:
K
8)8)Y)&
>
, we consider the following estimation
'2 . :
> ?@ ACB7DFEHG
W PX + 8+ Z, [ )
(3)
IKJ , U
> >
The goal of this paper is to show that as
,
in probability under appropri-
Given training examples
method to approximate the optimal predictor
ate regularity conditions. Furthermore, we obtain an estimate on the rate of convergence.
Consequences of this result in some specific learning formulations are examined.
2 Convergence of the estimated predictor
+ + ,
2 EHD I I ?
+ +8, R
I
W 2
2 *
=
*
*
*
2 *
2 *
*
It is clear that * can be regarded as a representing feature vector of in * . This representation can be computed as follows. Let + @MA B7D @-, I " , then it is not difficult to
/ + * +, and + ?+ * +, . It follows that * ?1
see that + .* 0
+ 2 + . Note that this
method of computing * is not important for the purpose of this paper.
Since * can now be considered as a linear functional using the feature space
representation
* of , we can use the idea from [6] to analyze the convergence behavior
>
of in * . Following
[6], using the linear representation of , we differentiate (2) at
>
the optimal solution , which leads to the following first order condition:
M ! 4 3 > *
5 * W P > TR
(4)
^
]
^
]
]
where 3
C_8 is the derivative of
C_ with respect to if is smooth; it denotes a
subgradient (see [4]) otherwise. Since we have assumed that ^]
_ is a convex function
of ] , we know that ^] -
_W ^] ] 8 3 ^] K
C_ 9 ^]
_ . This implies the following
Z
Z
inequality:
>
>
>
>
>
W
3
9
(
Assume that input belongs to a set . We make the reasonable assumption that is point
wise continuous under the
topology: ,
where
is in the sense that
. This assumption is equivalent to ! #"%$
'&
'( ) . The condition implies that each data point can be regarded as a bounded
linear functional * on
such that
: *
. Since a Hilbert space
is
self-dual, we can represent * by an element in . For notational simplicity, we shall
defined as *
for all
, where denotes the inner product of .
let *
which is equivalent to:
> (
& W PX > Z W
>
>
U 3 (
9 > (
W PX > Z )
Note that we have used Z to denote
we have
>
(
W PX > Z Q
> W4P > M > > [ W PX 6> > Z
+ + Z,
. Also note that by the definition of
> (
W PX > Z )
> ,
Therefore by comparing the above two inequalities, we obtain:
>
>
U 43 (
*
9 + 3 > (
5* W4P > + ,
PX > > Z9
X
>+ 6 > &+8, 9 +
PX
P+
This implies that
> W P > M > > [
+ 6> > + , )
3 >
* W P > +,
>
3
* ! 3 > (
5* + , )
(5)
> > >
(
Note that the last equality follows from the first order condition (4). This is the only place
the condition is used. In (5), we have already bounded the convergence of to in terms
of the convergence of the empirical expectation of a random vector 3
* to its
mean. The latter is often easier to estimate. For example, if its variance can be bounded,
then we may use the Chebyshev inequality to obtain a probability bound. In this paper,
we are interested in obtaining an exponential probability bound. In order to do so, similar
to the analysis in [6], we use the following form of concentration inequality which can be
found in [5], page 95:
R R + +,Q 9 X Z Q _ Z W + + , 9 Z
*
Z
3 > (
* MX ! 3 > (
5*
X
9 W 9 W4 7 X )
X
R R
Q
M ! 3 > (
Y+* +, 9 Z
X
+ > > +8, Q 9 , 5 P Z Z / ! _ Z W4P" )
_
>
X
Theorem 2.1 ([5]) Let be zero-mean
independent random vectors
in a Hilbert space
If there exists
such that for all natural numbers
:
/
Then for all
:
,
.
.
.
We may now use the following form of Jensen?s inequality to bound the moments of the
zero-mean random vector
:
From inequality (5) and Theorem 2.1, we immediately obtain the following bound:
Theorem 2.2 If there exists
such that for all natural numbers
. Then for all
:
:
Although Theorem 2.2 is quite general, the quantity and
on the right hand side of
the bound depend on the optimal predictor which requires to be estimated. In order to
obtain a bound that does not require any knowledge of the true distribution , we may
impose the following assumptions: both 3
and *
are bounded. Observe that
*
, we obtain the following result:
"
IKJ ,$# I
^ ]
_
+ +8,
Corollary 2.1 Assume that 2 , 5 IKJ ,%# I " 9'& . Also assume that the loss
function ^]
C_ satisfies the condition that 3 ^]
C_( 9 , then )* R :
+ > > + , Q+ 9 X, , P Z Z / ! & Z Z W & P" ()
+ +,
3 Generalization performance
We study some consequences of Corollary 2.1, which bounds the convergence rate of the
estimated predictor to the best predictor.
3.1 Regression
6
Z 6 Z ifif 6 9Q )
Z
It is clear that is continuous differentiable and 3
( 9 for all and
not hard to check that
and :
Z
K
3 -
9 X 8ZM)
Z
Z
Z
We consider the following type of Huber?s robust loss function:
7
(6)
. It is also
Using this inequality and (4), we obtain:
>
PX >
PX >
>
U M ! (
W + +Z, [ U M ! (
W + +8Z, [
\ ! > (
> (
3 > (
> > W
9 ! X > > Z W PX + > > +Z, )
> >
If we assume that + +8, 9 and IKJ ,$# I " 9 & , then
M ! >
9 M ! >
W PX + > + Z, + > + Z, W ZX
PX + > > + Z,
& Z W4P()
(7)
M ! > (
! > (
9 P" + > &+, W Z & X Z W P()
This gives the following inequality:
>
It is clear that the right-hand side of the above inequality does not depend on the unobserved
function . Using Corollary 2.1, we obtain the following bound:
+
X IKJ ,%! # I "
P Z Z / R & Z
9 & P
M ! > (
9 M ! > (
W4"P + > &+, W Z & X Z W
Theorem 3.1 Using loss function (6) in (3). Assume that
/ , with probability of at least
, 5
&
R
> 2 . *
9 & , then
Z , we have
P()
Theorem 3.1 compares the performance of the computed function with that of the optimal
predictor
in (1). This style of analysis has been extensively used in the
literature. For example, see [3] and references therein. In order to compare with
their
results, we can rewrite Theorem 3.1 in another form as: with probability of at least ,
M ! >
9 M ! >
W
HG
)
In [3], the authors employed a covering number analysis which led to a bound of the form
(for squared loss)
M ! > (
9 M ! > (
W
HG
HG
for finite dimensional problems. Note that the constant in their
depends on the pseudodimension, which can be infinity for problems considered in this paper. It is possible to employ their analysis using some covering number bounds for general Hilbert spaces. However, such an analysis would have led to a result of the following form for our problems:
M ! > (
9 M ! > (
W
G
G
()
It is also interesting to compare Theorem 3.1 with the leave-one-out analysis in [7]. The
generalization error averaged over all training examples for squared loss can be bounded
as
! >
9 W
P M ! > (
W4P+ > +Z, )
This result is not directly
> comparable with Theorem 3.1 since the right hand side includes
an extra term of P+ + Z, . Using the analysis in this paper, we may obtain a similar result
from (7) which leads to an average bound of the form:
! > (
9 M ! >
W P+ > + Z, &W
P Z ()
It is clear that the term resulted in our paper is not as good as
from [7].
However analysis in this paper leads to probability bounds while the leave-one-out analysis
in [7] only gives average bounds. It is also worth mentioning that it is possible to refine
the analysis presented in this section to obtain a probability bound which when averaged,
, rather than
in the current analysis.
gives a bound with the correct term of
However due to the space limitation, we shall skip this more elaborated derivation.
In addition to the above style bounds, it is also interesting to compare the generalization
performance of the computed function to the empirical error of the computed function.
Such results have occurred, for example, in [1]. In order to obtain a comparable result, we
may use a derivation similar to that of (7), together with the first order condition of (3) as
follows:
3 > *
5* W P >
R)
This leads to a bound
of the form:
> (
9
> (
W PX + > &+ Z,
+ > + Z, W X Z & Z W P )
X IKJ ,%! # I " 9 & , then
P
P Z Z / R & Z Z , we have
! >
> (
>
>
9 Z & Z W P W U M ! (
(
[ )
Combining the above inequality and (7), we obtain the following theorem:
+ 9 &
Theorem 3.2 Using loss function (6) in (3). Assume that
/ , with probability of at least
&
, 5
R
> (
M
!
U
>
>
Unlike Theorem 3.1, the bound given in Theorem 3.2 contains a term
which relies on the unknown optimal predictor . From Theorem 3.1,
we know that this term does not affect the performance of the estimated function when
>
[
>
compared with the performance of . In order for us to compare with the bound in [1]
obtained from an algorithmic stability point of view, we make the additional assumption
for all
. Note that this assumption is also required in [1]. Using
that
/
Hoeffding?s inequality, we obtain that with probability of at most , 5
,
> 9
! >
P Z Z ! R & Z Z
/
& )
> (
% P
M ! >
,
HG
> (
9 & Z W P ! R & Z Z HG
W Z
)
PZ
Together with Theorem 3.2, we have with probability of at least
This compares very favorably to the following bound in [1]:1
M ! >
> (
9
X
X
X
& P Z Z W Z P & W ! P& Z W X HG Z )
Z
3.2 Binary classification
20 <
QTR
1R
In binary classification, the output value
is a discrete variable. Given a continu, we consider the following prediction rule: predict
if
, and
ous model
otherwise. The classification error (we shall ignore the point
predict
,
which is assumed to occur rarely) is
(
R
if
if
9 R
R )
Unfortunately, this classification error function is not convex, which cannot be handled in
our formulation. In fact, even in many other popular methods, such as logistic regression
and support vector machines, some kind of convex formulations have to be employed. We
shall thus consider the following soft-margin SVM style loss as an illustration:
7DF@-,
R )
(8)
Note that the separable case of this loss was investigated in [6]. In this case, 3
denotes a subgradient
rather than gradient
since
is non-smooth: at
,
3
2 U
CR [ ; 3
when & and 3
7R when .
> >
Since , 9 and ,
9
, we know that if + +, 9 ,
Z
Z
then
M ! > (
9 ! > &
(
X
>
>
&
9
&
()
P Z Z ! R & Z
M ! >
Using the standard Hoeffding?s inequality, we have with probability of at most
/
, 5
,
1
&
>
&
W P / !
& )
In [1], there was a small error after equation (11). As a result, the original bound in their paper
was in a form equivalent to the one we cite here with
replaced by
.
>
When &
R , it is usually better to use a different (multi , 5 P Z form
plicative)
inequality, which implies that with probability of at most
Z / ! Rof& Z Hoeffding?s
,
X
>
>
M ! &
D @ , &
(
P Z Z / & Z ()
Together with Corollary 2.1, we obtain the following margin-percentile result:
IKJ ,%# I " 9
R + 9 & P
P Z Z / ! R & Z , we have
X
>
>
!
M ! (
9 &
W P / & )
We also have with probability of at least
, P Z Z / ! R & Z ,
X
X
M ! >
9 DF@-, > &
(
CP Z KZ / & Z8 ()
Theorem 3.3 Using loss function (8) in (3). Assume that
/ , with probability of at least
&
, 5
&
, then
We may obtain from Theorem 3.3 the following result: with probability of at least
! >
9
>
!
R G
PZ
&Z
&W
GX
,
)
It is interesting to compare this result with margin percentile style bounds from VC analysis. For example, Theorem
4.19 in [2] implies that there exists a constant such that with
probability of at least : for all we have
! > (
9
P
>
W
& PZ HG Z W HG
Z
!
R
& Z
)
We can see that if we assume that is small and the margin
is also small,
then the above bound with this choice of is inferior to the bound in Theorem 3.3. Clearly,
this implies that our analysis has some advantages over VC analysis due to the fact that we
directly analyze the numerical formulation of support vector classification.
4 Conclusion
In this paper, we have introduced a notion of the convergence of the estimated predictor
to the best underlying predictor for some learning problems in Hilbert spaces. This generalizes an earlier study in [6]. We derived generalization bounds for some regression and
classification problems. We have shown that results from our analysis compare favorably
with a number of earlier studies. This indicates that the concept introduced in this paper
can lead to valuable insights into certain numerical formulations of learning problems.
References
[1] Olivier Bousquet and Andr?e Elisseeff. Algorithmic stability and generalization performance. In Todd K. Leen, Thomas G. Dietterich, and Volker Tresp, editors, Advances
in Neural Information Processing Systems 13, pages 196?202. MIT Press, 2001.
[2] Nello Cristianini and John Shawe-Taylor. An Introduction to Support Vector Machines
and other Kernel-based Learning Methods. Cambridge University Press, 2000.
[3] Wee Sun Lee, Peter L. Bartlett, and Robert C. Williamson. The importance of convexity in learning with squared loss. IEEE Trans. Inform. Theory, 44(5):1974?1980,
1998.
[4] R. Tyrrell Rockafellar. Convex analysis. Princeton University Press, Princeton, NJ,
1970.
[5] Vadim Yurinsky. Sums and Gaussian vectors. Springer-Verlag, Berlin, 1995.
[6] Tong Zhang. Convergence of large margin separable linear classification. In Advances
in Neural Information Processing Systems 13, pages 357?363, 2001.
[7] Tong Zhang. A leave-one-out cross validation bound for kernel methods with applications in learning. In 14th Annual Conference on Computational Learning Theory,
pages 427?443, 2001.
| 2136 |@word dietterich:1 concept:1 skip:1 implies:6 multiplier:1 norm:1 true:2 equality:1 already:1 quantity:1 correct:1 concentration:1 vc:2 elisseeff:1 self:1 gradient:1 inferior:1 tr:1 covering:2 require:1 percentile:2 yorktown:1 moment:1 mx:1 berlin:1 contains:1 generalization:8 nello:1 reason:1 cp:1 c_:5 current:1 com:1 comparing:1 considered:2 wise:1 relationship:2 illustration:1 algorithmic:2 john:1 predict:3 difficult:1 numerical:2 functional:8 unfortunately:1 robert:1 favorably:2 negative:1 purpose:1 estimation:1 occurred:1 unknown:1 cambridge:1 finite:1 mit:1 clearly:1 gaussian:1 gx:1 shawe:1 rather:2 zhang:3 height:1 continu:1 cr:1 volker:1 corollary:4 introduced:2 derived:1 pair:1 required:1 notational:1 belongs:1 check:1 introduce:1 indicates:1 certain:1 verlag:1 huber:1 expected:1 inequality:13 behavior:1 sense:1 binary:2 watson:2 multi:1 dependent:1 ous:1 trans:1 usually:2 additional:1 below:1 impose:1 employed:2 interested:1 underlying:4 bounded:5 dual:1 classification:7 smooth:2 kind:1 technical:1 natural:2 cross:1 representing:1 unobserved:2 nj:1 prediction:1 regression:3 expectation:1 df:2 tresp:1 represent:1 control:1 employ:1 kernel:2 literature:1 randomly:1 wee:1 addition:2 resulted:1 todd:1 loss:12 replaced:1 consequence:2 qsr:1 interesting:3 limitation:1 extra:1 vadim:1 unlike:1 validation:1 investigate:1 therein:1 examined:1 editor:1 mentioning:1 ibm:2 hg:8 averaged:2 affect:1 last:1 topology:1 necessary:1 inner:1 idea:1 side:3 chebyshev:1 taylor:1 empirical:2 w4:2 handled:1 bartlett:1 z8:1 kz:1 soft:1 earlier:2 peter:1 author:1 cannot:1 introducing:1 approximate:1 subset:1 clear:4 equivalent:4 predictor:13 lagrangian:1 center:1 ignore:1 extensively:1 convex:5 assumed:2 simplicity:1 immediately:1 andr:1 rule:1 insight:1 continuous:2 regarded:2 estimated:6 lee:1 stability:2 robust:1 notion:2 discrete:1 shall:5 together:3 obtaining:1 squared:3 investigated:1 olivier:1 williamson:1 drawn:2 hoeffding:3 ifif:1 element:1 derivative:1 style:4 subgradient:2 yurinsky:1 sum:1 observed:1 rof:1 includes:1 rockafellar:1 tzhang:1 place:1 family:1 reasonable:1 sun:1 ny:1 depends:1 tong:3 qtr:1 valuable:1 view:1 exponential:1 analyze:2 comparable:2 convexity:1 start:1 bound:27 cristianini:1 theorem:19 refine:1 elaborated:1 depend:2 rewrite:2 annual:1 occur:1 specific:1 infinity:1 variance:1 appropri:1 jensen:1 pz:2 svm:1 bousquet:1 exists:3 importance:1 separable:2 derivation:2 worth:1 zx:1 px:15 margin:5 easier:1 inform:1 led:2 ate:1 quite:1 definition:1 otherwise:2 ikj:7 springer:1 cite:1 satisfies:1 popular:1 differentiate:1 advantage:1 differentiable:1 knowledge:1 equation:1 relies:1 ma:1 hilbert:7 goal:3 product:1 know:3 zm:1 combining:1 hard:1 generalizes:1 tyrrell:1 observe:1 formulation:7 leen:1 furthermore:1 convergence:10 regularity:1 hand:3 rarely:1 original:1 thomas:1 denotes:3 support:3 leave:3 latter:1 derive:1 logistic:1 quality:1 measured:1 princeton:2 pseudodimension:1 |
1,248 | 2,137 | Relative Density Nets: A New Way to
Combine Backpropagation with HMM's
Andrew D. Brown
Department of Computer Science
University of Toronto
Toronto, Canada M5S 3G4
[email protected]
Geoffrey E. Hinton
Gatsby Unit, UCL
London, UK WCIN 3AR
[email protected]
Abstract
Logistic units in the first hidden layer of a feedforward neural network compute the relative probability of a data point under two
Gaussians. This leads us to consider substituting other density
models. We present an architecture for performing discriminative
learning of Hidden Markov Models using a network of many small
HMM's. Experiments on speech data show it to be superior to the
standard method of discriminatively training HMM's.
1
Introduction
A standard way of performing classification using a generative model is to divide the
training cases into their respective classes and t hen train a set of class conditional
models. This unsupervised approach to classification is appealing for two reasons. It
is possible to reduce overfitting, because t he model learns the class-conditional input
densities P(xlc) rather t han the input -conditional class probabilities P(clx). Also,
provided that the model density is a good match to the underlying data density
then the decision provided by a probabilistic model is Bayes optimal. The problem
with this unsupervised approach to using probabilistic models for classification is
that, for reasons of computational efficiency and analytical convenience, very simple
generative models are typically used and the optimality of the procedure no longer
holds. For this reason it is usually advantageous to train a classifier discriminatively.
In this paper we will look specifically at the problem of learning HMM 's for classifying speech sequences. It is an application area where the assumption that the HMM
is the correct generative model for the data is inaccurate and discriminative methods
of training have been successful. The first section will give an overview of current
methods of discriminatively training HMM classifiers. We will then introduce a new
type of multi-layer backpropagation network which takes better advantage of the
HMM 's for discrimination. Finally, we present some simulations comparing the two
methods.
19 ' S1
c1
1
1
1
="
[tn] [tn][t n] HMM 's
\V
Sequence
Figure 1: An Alphanet with one HMM per class. Each computes a score for the
sequence and this feeds into a softmax output layer.
2
Alphanets and Discriminative Learning
The unsupervised way of using an HMM for classifying a collection of sequences is to
use the Baum-Welch algorithm [1] to fit one HMM per class. Then new sequences
are classified by computing the probability of a sequence under each model and
assigning it to the one with the highest probability. Speech recognition is one of the
commonest applications of HMM 's, but unfortunately an HMM is a poor model of
the speech production process. For this reason speech researchers have looked at the
possibility of improving the performance of an HMM classifier by using information
from negative examples - examples drawn from classes other than the one which
the HMM was meant to model. One way of doing this is to compute the mutual
information between the class label and the data under the HMM density, and
maximize that objective function [2].
It was later shown that this procedure could be viewed as a type of neural network
(see Figure 1) in which the inputs to the network are the log-probability scores
C(Xl:TIH) of the sequence under hidden Markov model H [3]. In such a model
there is one HMM per class, and the output is a softmax non-linearity:
(1)
Training this model by maximizing the log probability of correct classification leads
to a classifier which will perform better than an equivalent HMM model trained
solely in a unsupervised manner. Such an architecture has been termed an "AIphanet" because it may be implemented as a recurrent neural network which mimics
the forward pass of the forward-backward algorithm.l
3
Backpropagation Networks as Density Comparators
A multi-layer feedforward network is usually thought of as a flexible non-linear
regression model, but if it uses the logistic function non-linearity in the hidden
layer, there is an interesting interpretation of the operation performed by each
hidden unit. Given a mixture of two Gaussians where we know the component
priors P(9) and the component densities P(xl9) then the posterior probability that
Gaussian, 90 , generated an observation x , is a logistic function whose argument is
the negative log-odds of the two classes [4] . This can clearly be seen by rearranging
lThe results of the forward pass are the probabilities of the hidden states conditioned
on the past observations, or "alphas" in standard HMM terminology.
the expression for the posterior:
P(xI9o)P(Qo)
P(xI9o)P(Qo) + P(xI9d P (Qd
P(Qolx)
1
1 + exp {-log
P(x IQo) P(x lQd
log
(2)
P(Qo) }
P(Ql)
If the class conditional densities in question are multivariate Gaussians
P(xI9k) =
121f~1-~ exp {-~(x -
with equal covariance matrices,
written in this familiar form:
~,
Pk)T ~-l(X - Pk)}
(3)
then the posterior class probability may be
1
P(Qo Ix) = -l-+-e-xp-{-=---(:-x=Tw-+-b---:-)
(4)
where,
w
(5)
b
(6)
Thus, the multi-layer perceptron can be viewed as computing pairwise posteriors
between Gaussians in the input space, and then combining these in the output layer
to compute a decision.
4
A New Kind of Discriminative Net
This view of a feedforward network suggests variations in which other kinds of
density models are used in place of Gaussians in the input space. In particular,
instead of performing pairwise comparisons between Gaussians, the units in the
first hidden layer can perform pairwise comparisons between the densities of an
input sequence under M different HMM's. For a given sequence the log-probability
of a sequence under each HMM is computed and the difference in log-probability
is used as input to the logistic hidden unit. 2 This is equivalent to computing the
posterior responsibilities of a mixture of two HMM's with equal prior probabilities.
In order to maximally leverage the information captured by the HMM's we use (~)
hidden units so that all possible pairs are included. The output of a hidden unit h
is given by
(7)
where we have used (mn) as an index over the set, (~) , of all unordered pairs of
the HMM's. The results of this hidden layer computation are then combined using
a fully connected layer of free weights, W, and finally passed through a soft max
function to make the final decision.
ak
=
L
W(m ,n)kh(mn)
(8)
(mn) E (~)
(9)
2We take the time averaged log-probability so that the scale of the inputs is independent
of the length of the sequence.
Density
Comparator
Units
Figure 2: A multi-layer density net with HMM's in the input layer. The hidden
layer units perform all pairwise comparisons between the HMM 's.
where we have used u(?) as shorthand for the logistic function, and Pk is the value
of the kth output unit. The resulting architecture is shown in figure 2. Because
each unit in the hidden layer takes as input the difference in log-probability of two
HMM 's, this can be thought of as a fixed layer of weights connecting each hidden
unit to a pair of HMM's with weights of ?l.
In contrast to the Alphanet , which allocates one HMM to model each class, this network does not require a one-to-one alignment between models and classes and it gets
maximum discriminative benefit from the HMM's by comparing all pairs. Another
benefit of this architecture is that it allows us to use more HMM's than there are
classes. The unsupervised approach to training HMM classifiers is problematic because it depends on the assumption that a single HMM is a good model of the data
and, in the case of speech, this is a poor assumption. Training the classifier discriminatively alleviated this drawback and the multi-layer classifier goes even further in
this direction by allowing many HMM's to be used to learn the decision boundaries
between the classes. The intuition here is that many small HMM's can be a far
more efficient way to characterize sequences than one big HMM. When many small
HMM's cooperate to generate sequences, the mutual information between different
parts of generated sequences scales linearly with the number of HMM's and only
logarithmically with the number of hidden nodes in each HMM [5].
5
Derivative Updates for a Relative Density Network
The learning algorithm for an RDN is just the backpropagation algorithm applied
to the network architecture as defined in equations 7,8 and 9. The output layer is
a distribution over class memberships of data point Xl:T, and this is parameterized
as a softmax function. We minimize the cross-entropy loss function:
K
f =
2: tk logpk
(10)
k= l
where Pk is the value of the kth output unit and tk is an indicator variable which is
equal to 1 if k is the true class. Taking derivatives of this expression with respect
to the inputs of the output units yields
of
- = t k - Pk
oak
(11)
O?
o?
Oak
(12)
- , - - - - = (tk - Pk)h(mn)
OW(mn) ,k
oak OW(mn) ,k
The derivative of the output of the (mn)th hidden unit with respect to the output
of ith HMM, ?i, is
oh(mn)
(13)
~ = U(?m - ?n)(l - U(?m - ?n))(bim - bin)
where (bim - bin) is an indicator which equals +1 if i = m, -1 if i = n and zero
otherwise. This derivative can be chained with the the derivatives backpropagated
from the output to the hidden layer.
For the final step of the backpropagation procedure we need the derivative of the
log-likelihood of each HMM with respect to its parameters. In the experiments we
use HMM 's with a single, axis-aligned, Gaussian output density per state. We use
the following notation for the parameters:
?
?
?
?
?
A: aij is the transition probability from state i to state j
II: 7ri is the initial state prior
f./,i: mean vector for state i
Vi: vector of variances for state i
1-l: set of HMM parameters {A , II, f./" v}
We also use the variable St to represent the state of the HMM at time t. We make
use of the property of all latent variable density models that the derivative of the
log-likelihood is equal to the expected derivative of the joint log-likelihood under
the posterior distribution. For an HMM this means that:
O?(Xl:TI1-l)
'"
0
o1-l i
= ~ P(Sl:Tlxl:T' 1-l) o1-l i log P(Xl:T' Sl:TI1-l)
(14)
Sl:T
The joint likelihood of an HMM is:
(logP(Xl:T ' Sl:TI1-l)) =
T
L(b81 ,i)log 7ri
+ LL(b "jb
8
8
,_1 ,i)log aij
+
i,j
t=2
~ ~(b8" i) [-~ ~IOgVi'd ~ ~(Xt'd -
f./,i,d) 2 /Vi,d]
-
+ canst
(15)
where (-) denotes expectations under the posterior distribution and (b 8 , ,i) and
(b 8 , ,jb8 '_1 ,i) are the expected state occupancies and transitions under this distribution. All the necessary expectations are computed by the forward backward algorithm. We could take derivatives with respect to this functional directly, but that would require doing constrained gradient descent on the probInstead, we reparameterize the model using a
abilities and the variances.
softmax basis for probability vectors and an exponential basis for the variance parameters.
This choice of basis allows us to do unconstrained optimization in the new basis.
The new parameters are defined as follows:
. _
a' J -
exp(e;; ?)
(e (a? ) ,
2:
JI
exp
1JI
. _
7r, -
exp(e; ~?)
2:
if
exp
(e i(~?)'
.
_
(v)
V"d - exp(Oi,d )
This results in the following derivatives:
O?(Xl :T 11-l)
oO(a)
'J
T
L
t= 2
[(b 8 , ,jb 8 '_1 ,i) - (b 8 '_1 ,i)aij ]
(16)
8?(Xl:T 11?)
80(7r)
?
8?(Xl:T 11?)
8f..li,d
8?(Xl:T 11?)
80(v)
.,d
(8
S1
,i) -
(17)
1fi
T
l)8 st ,i)(Xt,d -
(18)
f..li ,d)/Vi ,d
t= l
1 T
2"l)8st ,i)
[(Xt ,d - f..li ,d)2/Vi ,d -
IJ
(19)
t= l
When chained with the error signal backpropagated from the output, these derivatives give us the direction in which to move the parameters of each HMM in order
to increase the log probability of the correct classification of the sequence.
6
Experiments
To evaluate the relative merits of the RDN, we compared it against an Alphanet
on a speaker identification task. The data was taken from the CSLU 'Speaker
Recognition' corpus. It consisted of 12 speakers uttering phrases consisting of 6
different sequences of connected digits recorded multiple times (48) over the course
of 12 recording sessions. The data was pre-emphasized and Fourier transformed
in 32ms frames at a frame rate of lOms. It was then filtered using 24 bandpass,
mel-frequency scaled filters. The log magnitude filter response was then used as the
feature vector for the HMM's. This pre-processing reduced the data dimensionality
while retaining its spectral structure.
While mel-cepstral coefficients are typically recommended for use with axis-aligned
Gaussians, they destroy the spectral structure of the data, and we would like to
allow for the possibility that of the many HMM's some of them will specialize on
particular sub-bands of the frequency domain. They can do this by treating the
variance as a measure of the importance of a particular frequency band - using
large variances for unimportant bands, and small ones for bands to which they pay
particular attention.
We compared the RDN with an Alphanet and three other models which were implemented as controls. The first of these was a network with a similar architecture
to the RDN (as shown in figure 2), except that instead of fixed connections of ?1,
the hidden units have a set of adaptable weights to all M of the HMM's. We refer
to this network as a comparative density net (CDN). A second control experiment
used an architecture similar to a CDN without the hidden layer, i.e. there is a single
layer of adaptable weights directly connecting the HMM's with the softmax output
units. We label this architecture a CDN-l. The CDN-l differs from the Alphanet
in that each softmax output unit has adaptable connections to the HMM's and we
can vary the number of HMM's, whereas the Alphanet has just one HMM per class
directly connected to each softmax output unit. Finally, we implemented a version
of a network similar to an Alphanet, but using a mixture of Gaussians as the input density model. The point of this comparison was to see if the HMM actually
achieves a benefit from modelling the temporal aspects of the speaker recognition
task.
In each experiment an RDN constructed out of a set of, M, 4-state HMM's was
compared to the four other networks all matched to have the same number of free
parameters, except for the MoGnet. In the case of the MoGnet, we used the same
number of Gaussian mixture models as HMM's in the Alphanet, each with the
same number of hidden states. Thus, it has fewer parameters, because it is lacking
the transition probabilities of the HMM. We ran the experiment four times with
a)
0.95
~
~
0.9
b)
~
0.95
e
E=:l
~
0.9
e =
0.85
0.85
0.8
0.8
0.75
0.75
8
0.7
0.65
0
0.7
0.6
0.6
0.55
Alphanet
MaGnet
CDN
EJ
0.65
0.55
RDN
B
CDN-1
RDN
Alphanet
Architecture
C)
$
D
~
d)
e
~
0.9
8
*
c
0
CDN-1
~
*0.8
?in
gj
Ci
gj
CiO.5
B
MeG net
Architecture
CDN
U
gO.7
~
~0.6
?in
Alphanet
~
a:
~
~O.8
RDN
CDN
~
a:
0.6
MaGnet
Architecture
CDN-1
8
0.4
0.3
RDN
Alphanet
MeGnet
CDN
CDN-1
Architecture
Figure 3: Results of the experiments for an RDN with (a) 12, (b) 16, (c) 20 and
(d) 24 HMM's.
values of M of 12, 16, 20 and 24. For the Alphanet and MoGnet we varied the
number of states in the HMM's and the Gaussian mixtures, respectively. For the
CDN model we used the same number of 4-state HMM's as the RDN and varied
the number of units in the hidden layer of the network. Since the CDN-1 network
has no hidden units, we used the same number of HMM's as the RDN and varied
the number of states in the HMM. The experiments were repeated 10 times with
different training-test set splits. All the models were trained using 90 iterations of
a conjugate gradient optimization procedure [6] .
7
Results
The boxplot in figure 3 shows the results of the classification performance on the
10 runs in each of the 4 experiments. Comparing the Alphanet and the RDN we
see that the RDN consistently outperforms the Alphanet. In all four experiments
the difference in their performance under a paired t-test was significant at the level
p < 0.01. This indicates that given a classification network with a fixed number of
parameters, there is an advantage to using many small HMM 's and using all the
pairwise information about an observed sequence, as opposed to using a network
with a single large HMM per class.
In the third experiment involving the MoGnet we see that its performance is comparable to that of the Alphanet. This suggests that the HMM's ability to model the
temporal structure of the data is not really necessary for the speaker classification
task as we have set it Up.3 Nevertheless, the performance of both the Alphanet and
3If we had done text-dependent speaker identification, instead of multiple digit phrases
the MoGnet is less than the RDN.
Unfortunately the CDN and CDN-l networks perform much worse than we expected. While we expected these models to perform similarly to the RDN, it seems
that the optimization procedure takes much longer with these models. This is probably because the small initial weights from the HMM's to the next layer severely
attenuate the backpropagated error derivatives that are used to train the HMM's.
As a result the CDN networks do not converge properly in the time allowed.
8
Conclusions
We have introduced relative density networks, and shown that this method of discriminatively learning many small density models in place of a single density model
per class has benefits in classification performance. In addition, there may be a
small speed benefit to using many smaller HMM 's compared to a few big ones.
Computing the probability of a sequence under an HMM is order O(TK 2 ), where T
is the length of the sequence and K is the number of hidden states in the network.
Thus, smaller HMM 's can be evaluated faster. However, this is somewhat counterbalanced by the quadratic growth in the size of the hidden layer as M increases.
Acknowledgments
We would like to thank John Bridle, Chris Williams, Radford Neal, Sam Roweis ,
Zoubin Ghahramani, and the anonymous reviewers for helpful comments.
References
[1] L. E. Baum, T. Petrie, G. Soules, and N. Weiss, "A maximization technique
occurring in the statistical analysis of probabilistic functions of Markov chains,"
The Annals of Mathematical Statistics, vol. 41, no. 1, pp. 164-171, 1970.
[2] 1. R. Bahl, P. F. Brown, P. V. de Souza, and R. 1. Mercer, "Maximum mutual information of hidden Markov model parameters for speech recognition,"
in Proceeding of the IEEE International Conference on Acoustics, Speech and
Signal Processing, pp. 49- 53, 1986.
[3] J. Bridle, "Training stochastic model recognition algorithms as networks can
lead to maximum mutual information estimation of parameters," in Advances in
Neural Information Processing Systems (D. Touretzky, ed.), vol. 2, (San Mateo,
CA), pp. 211- 217, Morgan Kaufmann, 1990.
[4] M. I. Jordan, "Why the logistic function? A tutorial discussion on probabilities
and neural networks," Tech. Rep. Computational Cognitive Science, Technical
Report 9503, Massachusetts Institute of Technology, August 1995.
[5] A. D. Brown and G. E. Hinton, "Products of hidden Markov models," in Proceedings of Artificial Intelligence and Statistics 2001 (T. Jaakkola and T. Richardson, eds.), pp. 3- 11, Morgan Kaufmann, 2001.
[6] C. E. Rasmussen, Evaluation of Gaussian Processes and other Methods for NonLinear Regression. PhD thesis, University of Toronto, 1996. Matlab conjugate
gradient code available from http ://www .gatsby.ucl.ac.uk/~edward/code/.
then this might have made a difference.
| 2137 |@word version:1 advantageous:1 seems:1 simulation:1 covariance:1 initial:2 score:2 past:1 outperforms:1 current:1 comparing:3 soules:1 assigning:1 written:1 john:1 treating:1 update:1 discrimination:1 generative:3 fewer:1 intelligence:1 ith:1 filtered:1 node:1 toronto:3 oak:3 mathematical:1 constructed:1 shorthand:1 specialize:1 combine:1 manner:1 introduce:1 g4:1 pairwise:5 expected:4 multi:5 provided:2 underlying:1 linearity:2 notation:1 matched:1 kind:2 temporal:2 growth:1 classifier:7 scaled:1 uk:3 control:2 unit:21 severely:1 ak:1 solely:1 might:1 mateo:1 suggests:2 bim:2 averaged:1 acknowledgment:1 differs:1 backpropagation:5 digit:2 procedure:5 area:1 thought:2 alleviated:1 pre:2 zoubin:1 get:1 convenience:1 www:1 equivalent:2 reviewer:1 baum:2 maximizing:1 go:2 attention:1 williams:1 welch:1 oh:1 variation:1 annals:1 us:1 logarithmically:1 recognition:5 observed:1 connected:3 highest:1 ran:1 intuition:1 chained:2 trained:2 efficiency:1 basis:4 joint:2 train:3 london:1 artificial:1 whose:1 otherwise:1 ability:2 statistic:2 richardson:1 final:2 sequence:19 advantage:2 net:5 analytical:1 ucl:3 product:1 aligned:2 combining:1 roweis:1 kh:1 wcin:1 comparative:1 tk:4 oo:1 andrew:1 ac:2 recurrent:1 ij:1 edward:1 implemented:3 c:1 qd:1 direction:2 drawback:1 correct:3 filter:2 stochastic:1 bin:2 require:2 really:1 anonymous:1 hold:1 exp:7 substituting:1 vary:1 achieves:1 estimation:1 label:2 alphanet:17 clearly:1 gaussian:5 rather:1 ej:1 jaakkola:1 canst:1 properly:1 consistently:1 modelling:1 likelihood:4 indicates:1 tech:1 contrast:1 helpful:1 dependent:1 membership:1 inaccurate:1 typically:2 cslu:1 hidden:26 transformed:1 classification:9 flexible:1 retaining:1 constrained:1 softmax:7 mutual:4 equal:5 look:1 unsupervised:5 comparators:1 mimic:1 jb:2 report:1 few:1 familiar:1 consisting:1 possibility:2 evaluation:1 alignment:1 mixture:5 chain:1 andy:1 necessary:2 respective:1 allocates:1 divide:1 soft:1 ar:1 logp:1 maximization:1 phrase:2 successful:1 characterize:1 combined:1 st:3 density:21 international:1 probabilistic:3 connecting:2 b8:1 thesis:1 recorded:1 opposed:1 worse:1 cognitive:1 derivative:12 li:3 de:1 unordered:1 coefficient:1 xlc:1 depends:1 vi:4 later:1 performed:1 view:1 responsibility:1 doing:2 bayes:1 minimize:1 oi:1 cio:1 variance:5 kaufmann:2 yield:1 identification:2 researcher:1 m5s:1 classified:1 touretzky:1 ed:2 against:1 frequency:3 pp:4 bridle:2 massachusetts:1 dimensionality:1 actually:1 adaptable:3 feed:1 response:1 maximally:1 wei:1 done:1 evaluated:1 just:2 qo:4 nonlinear:1 logistic:6 bahl:1 brown:3 true:1 consisted:1 neal:1 ll:1 speaker:6 mel:2 m:1 tn:2 cooperate:1 fi:1 petrie:1 superior:1 functional:1 ji:2 overview:1 he:1 interpretation:1 refer:1 significant:1 attenuate:1 unconstrained:1 session:1 similarly:1 had:1 han:1 longer:2 gj:2 posterior:7 multivariate:1 termed:1 rep:1 seen:1 captured:1 morgan:2 somewhat:1 converge:1 maximize:1 recommended:1 signal:2 ii:2 multiple:2 technical:1 match:1 faster:1 cross:1 clx:1 paired:1 involving:1 regression:2 expectation:2 iteration:1 represent:1 c1:1 whereas:1 addition:1 probably:1 comment:1 recording:1 jordan:1 odds:1 leverage:1 feedforward:3 split:1 fit:1 counterbalanced:1 architecture:12 reduce:1 ti1:3 expression:2 passed:1 speech:8 matlab:1 unimportant:1 backpropagated:3 band:4 reduced:1 generate:1 http:1 sl:4 problematic:1 tutorial:1 per:7 vol:2 four:3 terminology:1 nevertheless:1 drawn:1 backward:2 destroy:1 run:1 parameterized:1 place:2 decision:4 comparable:1 layer:23 pay:1 quadratic:1 ri:2 boxplot:1 fourier:1 aspect:1 argument:1 optimality:1 reparameterize:1 speed:1 performing:3 department:1 poor:2 conjugate:2 smaller:2 sam:1 appealing:1 tw:1 s1:2 taken:1 equation:1 know:1 merit:1 available:1 gaussians:8 operation:1 spectral:2 denotes:1 ghahramani:1 objective:1 move:1 question:1 looked:1 ow:2 kth:2 gradient:3 thank:1 hmm:71 chris:1 lthe:1 reason:4 meg:1 length:2 o1:2 index:1 code:2 ql:1 unfortunately:2 negative:2 perform:5 allowing:1 observation:2 markov:5 descent:1 hinton:3 frame:2 varied:3 august:1 canada:1 souza:1 tih:1 introduced:1 pair:4 connection:2 acoustic:1 usually:2 max:1 indicator:2 mn:8 occupancy:1 technology:1 axis:2 text:1 prior:3 hen:1 relative:5 lacking:1 fully:1 loss:1 discriminatively:5 interesting:1 geoffrey:1 cdn:17 xp:1 mercer:1 classifying:2 production:1 course:1 free:2 loms:1 rasmussen:1 aij:3 allow:1 perceptron:1 institute:1 taking:1 cepstral:1 benefit:5 boundary:1 transition:3 computes:1 forward:4 collection:1 uttering:1 san:1 made:1 far:1 alpha:1 overfitting:1 corpus:1 discriminative:5 latent:1 why:1 learn:1 ca:2 rearranging:1 improving:1 domain:1 pk:6 linearly:1 big:2 alphanets:1 repeated:1 allowed:1 gatsby:3 sub:1 bandpass:1 exponential:1 xl:9 third:1 learns:1 ix:1 rdn:16 xt:3 emphasized:1 utoronto:1 importance:1 ci:1 phd:1 magnitude:1 conditioned:1 occurring:1 entropy:1 radford:1 conditional:4 comparator:1 viewed:2 included:1 specifically:1 except:2 pas:2 meant:1 evaluate:1 magnet:2 |
1,249 | 2,138 | Scaling of Probability-Based Optimization
Algorithms
J. L. Shapiro
Department of Computer Science University of Manchester
Manchester, M13 9PL U.K. [email protected]
Abstract
Population-based Incremental Learning is shown require very sensitive scaling of its learning rate. The learning rate must scale with
the system size in a problem-dependent way. This is shown in two
problems: the needle-in-a haystack, in which the learning rate must
vanish exponentially in the system size, and in a smooth function
in which the learning rate must vanish like the square root of the
system size. Two methods are proposed for removing this sensitivity. A learning dynamics which obeys detailed balance is shown to
give consistent performance over the entire range of learning rates.
An analog of mutation is shown to require a learning rate which
scales as the inverse system size, but is problem independent.
1
Introduction
There has been much recent work using probability models to search in optimization
problems. The probability model generates candidate solutions to the optimization
problem. It is updated so that the solutions generated should improve over time.
Usually, the probability model is a parameterized graphical model , and updating
the model involves changing the parameters and possibly the structure of the model.
The general scheme works as follows,
? Initialize the model to some prior (e.g. a uniform distribution);
? Repeat
- Sampling step: generate a data set by sampling from the probability
model;
- Testing step: test the data as solutions to the problem;
- Selection step: create a improved data set by selecting the better
solutions and removing the worse ones;
- Learning step: create a new probability model from the old model
and the improved data set (e.g. as a mixture of the old model and the
most likely model given the improved data set);
? until (stopping criterion met)
Different algorithms are largely distinguished by the class of probability models
used. For reviews of the approach including the different graphical models which
have been used, see [3, 6]. These algorithms have been called Estimation of Distribution Algorithms (EDA); I will use that term here.
EDAs are related to genetic algorithms; instead of evolving a population, a generative model which produces the population at each generation is evolved. A
motivation for using EDAs instead of GAs is that is that in EDAs the structure of
the graphical model corresponds to the form of the crossover operator in GAs (in
the sense that a given graph will produce data whose probability will not change
much under a particular crossover operator). If the EDA can learn the structure of
the graph, it removes the need to set the crossover operator by hand (but see [2]
for evidence against this).
In this paper, a very simple EDA is considered on very simple problems. It is shown
that the algorithm is extremely sensitive to the value of learning rate. The learning
rate must vanish with the system size in a problem dependent way, and for some
problems it has to vanish exponentially fast. Two correctives measures are considered: a new learning rule which obeys detailed balance in the space of parameters,
and an operator analogous to mutation which has been proposed previously.
2
The Standard PBIL Algorithm
The simplest example of a EDA is Population-based Incremental Learning (PBIL)
which was introduced by Baluja [1]. PBIL uses a probability model which is a
product of independent probabilities for each component of the binary search space.
Let Xi denote the ith component of X, an L-component binary vector which is a
state of the search space. The probability model is defined by the L-component
vector of parameters 'Y ~), where 'Yi(t ) denotes the probability that Xi = 1 at time
t.
The algorithm works as follows,
? Initialize 'Yi(O) = 1/2 for all i;
? Repeat
- Generate a population of N strings by sampling from the binomial
distribution defined by 1(t).
- Find the best string in the population x*.
- Update the parameters 'Yi(t + 1) = 'Yi(t) + a[xi - 'Yi (t)] for all i.
? until (stopping criterion met)
The algorithm has only two parameters, the size of the population N and the
learning parameter a.
3
3.1
The sensitivity of PBIL to the learning rate
PBIL on a flat landscape
The source of sensitivity of PBIL to the learning rate lies in its behavior on a flat
landscape. In this case all vectors are equally fit , so the "best" vector x* is a random
vector and its expected value is
(1)
(where (-) denotes the expectation operator) Thus, the parameters remain unchanged on average. In any individual run, however, the parameters converge
rapidly to one of the corners of the hypercube. As the parameters deviate from
1/2 they will move towards a corner of the hypercube. Then the population generated will be biased towards that corner, which will move the parameters closer
yet to that corner, etc. All of the corners of the hypercube are attractors which,
although never reached, are increasingly attractive with increasing proximity. Let
us call this phenomenon drift. (In population genetics, the term drift refers to the
loss of genetic diversity due to finite population sampling. It is in analogy to this
that the term is used here.)
Consider the average distance between the parameters and 1/2,
D(t) ==
1 (1
L 2: "2 - 'Yi (t)
?
)2
(2)
Solving this reveals that on average this converges to 1/4 with a characteristic time
T
= -1/ 10g(1 - 0: 2) ~ 1/0: 2 for 0: ~ O.
(3)
The rate of search on any other search space will have to compete with drift.
3.2
PBIL and the needle-in-the haystack problem
As a simple example of the interplay between drift and directed search, consider the
so-called needle-in-a-haystack problem. Here the fitness of all strings is 0 except for
one special string (the "needle") which has a fitness of 1. Assume it is the string
of all 1 'so It is shown here that PBIL will only find the needle if 0: is exponentially
small, and is inefficient at finding the needle when compared to random search.
rrf=1
'Yi(t).
Consider the probability of finding the needle at time t, denoted O(t) =
Consider times shorter than T where T is long enough that the needle may be
found multiple times, but 0:2T -+ 0 as L -+ 00. It will be shown for small 0: that
when the needle is not found (during drift), 0 decreases by an amount 0: 2LO/2,
whereas when the needle is found, 0 increases by the amount o:LO. Since initially,
the former happens at a rate 2L times greater than the latter, 0: must be less than
2 - (L - 1) for the system to move towards the hypercube corner near the optimum,
rather than towards a random corner.
When the needle is not found, the mean of O(t) is invariant, (O(t + 1)) = O(t).
However, this is misleading, because 0 is not a self-averaging quantity; its mean
is affected by exponentially unlikely events which have an exponentially big effect.
A more robust measure of the size of O(t) is the exponentiated mean of the log of
O(t) . This will be denoted by [0] == exp (log 0). This is the appropriate measure of
the central tendency of a distribution which is approximately log-normal [4], as is
expected of O(t) early in the dynamics, since the log of 0 is the sum of approximately
independent quantities.
The recursion for 0 expanded to second order in 0: obeys
[O(t
+ 1)] =
{
[O(t)] [1
[O(t)] [1
- 10:2 L] .
+ ~L + ~'0:2 L(L -
needle not found
1)] ;
needle found.
(4)
In these equations, 'Yi(t) has also been expanded around 1/2.
Since the needle will be found with probability O(t) and not found with probability
1 - O(t), the recursion averages to,
[O(t + 1)] = [O(t)] (1 -
~0:2 L) + [0(t)]2
[O:L -
~0:2 L(L + 1)] .
(5)
The second term actually averages to [D(t)] (D(t)) , but the difference between (D)
and [D] is of order 0:, and can be ignored.
Equation (5) has a stable fixed point at 0 and an unstable fixed point at 0:/2 +
O( 0: 2 L). If the initial value of D(O) is less than the unstable fixed point, D will
decay to zero. If D(O) is greater than the unstable fixed point, D will grow. The
initial value is D(O) = 2- ?, so the condition for the likelihood of finding the needle
to increase rather than decrease is 0: < 2-(?-1).
1.1 ,-----~-~--~-~--,_________,
120
a
Figure 1: Simulations on PBIL on needle-in-a-haystack problem for L = 8,10,11,12
(respectively 0, +, *, 6). The algorithm is run until no parameters are between 0.05
and 0.95, and averaged over 1000 runs. Left: Fitness of best population member at
convergence versus 0:. The non-robustness of the algorithm is clear; as L increases,
0: must be very finely set to a very small value to find the optimum. Right: As
previous, but with 0: scaled by 2?. The data approximately collapses, which shows
that as L increases, 0: must decrease like 2-? to get the same performance.
Figure 1 shows simulations of PBIL on the needle-in-a-haystack problem. These
confirm the predictions made above, the optimum is found only if 0: is smaller than
a constant times 2?. The algorithm is inefficient because it requires such small 0:;
convergence to the optimum scales like 4?. This is because the rate of convergence
to the optimum goes like Do:, both of which are 0(2-?).
3.3
PBIL and functions of unitation
One might think that the needle-in-the-haystack problem is hard in a special way,
and results on this problem are not relevant to other problems. This is not be true,
because even smooth functions have fiat subspaces in high dimensions. To see this,
consider any continuous, monotonic function of unit at ion u, where u =
L~ Xi , the
number of 1 's in the vector. Assume the the optimum occurs when all components
are l.
t
The parameters 1 can be decomposed into components parallel and perpendicular
to the optimum. Movement along the perpendicular direction is neutral, Only
movement towards or away from the optimum changes the fitness. The random
strings generated at the start of the algorithm are almost entirely perpendicular to
the global optimum, projecting only an amount of order 1/..JL towards the optimum.
Thus, the situation is like that of the needle-in-a-haystack problem. The perpendicular direction is fiat, so there is convergence towards an arbitrary hypercube corner
with a drift rate,
TJ.. '"
a?
(6)
from equation (3). Movement towards the global optimum occurs at a rate,
a
Til '" VL?
(7)
Thus, a must be small compared to l/VL for movement towards the global optimum
to win.
A rough argument can be used to show how the fitness in the final population
depends on a. Making use of the fact that when N random variables are drawn
from a Gaussian distribution with mean m and variance u 2 , the expected largest
value drawn is m + J2u 2 10g(N) for large N (see, for example, [7]) , the Gaussian
approximation to the binomial distribution, and approximating the expectation of
the square root as the square root of the expectation yields,
(u(t
+ 1)) =
(u(t))
+ aJ2 (v(t)) 10g(N),
(8)
-b
where v(t) is the variance in probability distribution, v(t) =
L i Ii (t)[l - li(t)].
Assuming that the convergence of the variance is primarily due to the convergence
on the flat subspace, this can be solved as,
(u(oo))
~
Jlog(N)
1
"2 + aV'iL .
(9)
The equation must break down when the fitness approaches one, which is where the
Gaussian approximation to the binomial breaks down.
0.9
0 .9
0 .8
0 .8
0.7
0 .7
0 .6
~
~ 0.6
~ 0.5
u..
I
NO .5
'" 0.4
0.4
0.3
0 .3
0 .2
0 .2
0.1
0
0
0.2
0.4
0.6
0.8
20
a
Figure 2: Simulations on PBIL on the unitation function for L = 16,32,64,128,256
(respectively D , 0, +, *, 6) . The algorithm is run until all parameters are closer to
1 or 0 than 0.05, and averaged over 100 runs. Left: Fitness of best population
member at convergence versus a. The fitness is scaled so that the global optimum
has fitness 1 and the expected fitness of a random string is O. As L increases, a
must be set to a decreasing value to find the optimum. Right: As previous, but
with a scaled by VL. The data approximately collapses, which shows t hat as L
increases, a must decrease like VL to get the same performance. The smooth curve
shows equation (9).
Simulations of PBIL on the unitation function confirm these predictions. PBIL fails
to converge to the global optimum unless a is small compared to l/VL. Figure 2
shows the scaling of fitness at convergence with aVL, and compares simulations
with equation (9).
4
Corrective 1 -
Detailed Balance PBIL
One view of the problem is that it is due to the fact that the learning dynamics
does not obey detailed balance. Even on a flat space, the rate of movement of
the parameters "Yi away from 1/2 is greater than the movement back. It is wellknown that a Markov process on variables x will converge to a desired equilibrium
distribution 7r(x) if the transition probabilities obey the detailed balance conditions,
w(x'lx)7r(x) = w(xlx')7r(x'),
(10)
where w(x'lx) is the probability of generating x' from x. Thus, any search algorithm
searching on a flat space should have dynamics which obeys,
w(x'lx)
= w(xlx'),
(11)
and PEIL does not obey this. Perhaps the sensitive dependence on a would be
removed if it did.
There is a difficulty in modifying the dynamics of PBIL to satisfy detailed balance,
however. PEIL visits a set of points which varies from run to run, and (almost)
never revisits points. This can be fixed by constraining the parameters to lie on a
lattice. Then the dynamics can be altered to enforce detailed balance.
Define the allowed parameters in terms of a set of integers ni. The relationship
between them is.
I - ~(1 - a)ni, ni > 0;
{
(12)
"Yi =
!(1- a) lni l,
ni < 0;
ni = O.
2'
Learning dynamics now consists of incrementing and decrementing the n/s by 1;
when xi = 1(0) ni is incremented (decremented).
Transforming variables via equation (12), the uniform distribution in "Y becomes in
n,
P (n) = _a_(I_ a) lnl.
(13)
2-a
4.0.1
Detailed balance by rejection sampling
One of the easiest methods for sampling from a distribution is to use the rejection
method. In this , one has g(x'lx) as a proposal distribution; it is the probability of
proposing the value x' from x. Then, A(x'lx) is the probability of accepting this
change. Detailed balance condition becomes
g(x'lx)A(x'lx)7r(x) = g(xlx')A(xlx')7r(x') .
(14)
For example, the well-known Metropolis-Hasting algorithm has
A(x'lx) = min
(1, :~~}:(~}I~})'
(15)
The analogous equations for PEIL on the lattice are,
A(n
+ lin)
A(n-lln)
=
. [1- "Y (n+l)
]
mm
"Y(n)
(1 - a), 1
(16)
min[{~;(~~(1-a),I].
(17)
In applying the acceptance formula, each component is treated independently. Thus,
moves can be accepted on some components and not on others.
4.0.2
Results
Detailed Balance PBIL requires no special tuning of parameters, at least when
applied to the two problems of the opening sections. For the needle-in-a-haystack,
simulations were performed for 100 values of (): between 0 and 0.4 equally spaced for
L = 8,9,10,11,12; 1000 trials of each, population size 20, with the same convergence
criterion as before, simulation halts when all "Ii'S are less than 0.05 or greater than
0.95. On none of those simulations did the algorithm fail to contain the global
optimum in the final population.
For the function of unitation, Detailed Balance PBIL appears to always find the
optimum if run long enough. Stopping it when all parameters fell outside the range
(0.05,0.95), the algorithm did not always find the global optimum. It produced
an average fitness within 1% of the optimum for (): between 0.1 and 0.4 and L =
32, 64,128,256 over a 100 trials, but for learning rates below 0.1 and L = 256 the
average fitness fell as low as 4% below optimum. However, this is much improved
over standard PBIL (see figure 2) where the average fitness fell to 60% below the
optimum in that range.
5
Corrective 2 -
Probabilistic mutation
Another approach to control drift is to add an operator analogous to mutation in
GAs. Mutation has the property that when repeatedly applied, it converges to a
random data set. Muhlenbein [5] has proposed that the analogous operator ED As
estimates frequencies biased towards a random guess. Suppose ii is the fraction of
l's at site i. Then, the appropriate estimate of the probability of a 1 at site i is
ii + m
(18)
"Ii = 1 + 2m'
where m is a mutation-like parameter. This will be recognized as the maximum
posterior estimate of the binomial distribution using as the prior a ,a-distribution
with both parameters equal to mN + 1; the prior biases the estimate towards 1/2.
This can be applied to PBIL by using the following learning rule,
(
"Ii t
+
1)
"Ii(t)
=
+ (): [x; - "Ii (t)] + m
1 + 2m
.
(19)
With m = 0 it gives the usual PBIL rule; when repeatedly applied on a flat space
it converges to 1/2.
Unlike Detailed Balance PBIL, this approach does required special scaling of the
learning rate, but the scaling is more benign than in standard PBIL and is problem
independent. It is determined from three considerations. First, mutation must
be large enough to counteract the effects of drift towards random corners of the
hypercube. Thus, the fixed point of the average distance to 1/2, (D(t + 1)) defined
in equation (2) , must be sufficiently close to zero. Second, mutation must be small
enough that it does not interfere with movement towards the parameters near the
optimum when the optimum is found. Thus, the fixed point of equation (19) must be
sufficiently close to 0 or 1. Finally, a sample of size N sampled from the fixed point
distribution near the hypercube corner containing the optimum should contain the
optimum with a reasonable probability (say greater than 1 - e- 1 ). Putting these
considerations together yields,
logN
m
():
-L- ? -(): ? -.
4
(20)
5.1
Results
To satisfy the conditions in equation 20, the mutation rate was set to m ex: a 2 ,
and a was constrained to be smaller than log (N)/L. For the needle-in-a-haystack,
the algorithm behaved like Detailed Balance PElL. It never failed to find the optimum for the needle-in-a-haystack problems for the sizes given previously. For the
functions of unitation, no improvement over standard PBIL is expected, since the
scaling using mutation is worse, requiring a < 1/ L rather than a < 1/..fL. However, with tuning of the mutation rate, the range of a's with which the optimum
was always found could be increased over standard PBIL.
6
Conclusions
The learning rate of PBIL has to be very small for the algorithm to work, and
unpredictably so as it depends upon the problem size in a problem dependent way.
This was shown in two very simple examples. Detailed balance fixed the problem
dramatically in the two cases studied. Using detailed balance, the algorithm consistently finds the optimum over the entire range of learning rates. Mutation also fixed
the problem when the parameters were chosen to satisfy a problem-independent set
of inequalities.
The phenomenon studied here could hold in any EDA, because for any type of
model, the probability is high of generating a population which reinforces the move
just made. On the other hand, more complex models have many more parameters, and also have more sources of variability, so the issue may be less important.
It would be interesting to learn how important this sensitivity is in EDAs using
complex graphical models.
Of the proposed correctives, detailed balance will be more difficult to generalize to
models in which the structure is learned. It requires an understanding of algorithm's
dynamics on a flat space, which may be very difficult to find in those cases. The
mutation-type operator will easier to generalize, because it only requires a bias
towards a random distribution. However, the appropriate setting of the parameters
may be difficult to ascertain.
References
[1] S. Baluja. Population-based incremental learning: A method for integrating genetic
search based function optimization and competive learning. Technical Report CMUCS-94-163 , Computer Science Department , Carnegie Mellon University, 1994.
[2] A. Johnson and J. L. Shapiro. The importance of selection mechanisms in distribution
estimation algorithms. In Proceedings of the 5th International Conference on Artificial
Evolution AE01, 2001.
[3] P. Larraiiaga and J. A. Lozano. Estimation of Distribution Algorithms, A New Tool
for Evolutionary Computation. Kluwer Academic Publishers, 2001.
[4] Eckhard Limpert , Werner A. Stahel, and Markus Abbt . Log-normal distributions
across the sciences: Keys and clues. BioScience, 51(5):341-352, 2001.
[5] H. Miihlenbein. The equation for response to selection and its use for prediction.
Evolutionary Computation, 5(3):303- 346, 1997.
[6] M. Pelikan, D. E . Goldberg, and F. Lobo. A survey of optimization by building
and using probabilistic models. Technical report, University of Illinois at UrbanaChampaign, Illinois Genetic Algorithms Laboratory, 1999.
[7] Jonathan L. Shapiro and Adam Priigel-Bennett. Maximum entropy analysis of genetic
algorithm operators. Lecture Notes in Computer Science, 993:14- 24, 1995.
| 2138 |@word trial:2 simulation:8 pbil:26 initial:2 selecting:1 genetic:5 yet:1 must:15 benign:1 remove:1 update:1 generative:1 guess:1 ith:1 stahel:1 accepting:1 lx:8 along:1 consists:1 expected:5 behavior:1 decomposed:1 decreasing:1 increasing:1 becomes:2 easiest:1 evolved:1 string:7 proposing:1 finding:3 scaled:3 uk:1 control:1 unit:1 priigel:1 before:1 jls:1 approximately:4 might:1 studied:2 collapse:2 range:5 perpendicular:4 obeys:4 averaged:2 directed:1 testing:1 evolving:1 crossover:3 integrating:1 refers:1 get:2 needle:22 selection:3 operator:9 close:2 applying:1 go:1 independently:1 survey:1 rule:3 lni:1 population:17 searching:1 analogous:4 updated:1 suppose:1 us:1 goldberg:1 updating:1 solved:1 decrease:4 movement:7 removed:1 incremented:1 transforming:1 lobo:1 dynamic:8 solving:1 upon:1 aj2:1 corrective:4 fast:1 artificial:1 outside:1 whose:1 say:1 think:1 final:2 interplay:1 product:1 relevant:1 rapidly:1 manchester:2 convergence:9 optimum:28 produce:2 generating:2 incremental:3 converges:3 adam:1 oo:1 urbanachampaign:1 ac:1 c:1 involves:1 met:2 direction:2 modifying:1 require:2 pl:1 mm:1 proximity:1 around:1 considered:2 sufficiently:2 normal:2 exp:1 hold:1 equilibrium:1 early:1 estimation:3 sensitive:3 largest:1 create:2 tool:1 rough:1 gaussian:3 always:3 rather:3 improvement:1 consistently:1 likelihood:1 sense:1 dependent:3 stopping:3 vl:5 entire:2 unlikely:1 initially:1 issue:1 denoted:2 logn:1 constrained:1 special:4 initialize:2 equal:1 never:3 sampling:6 others:1 decremented:1 report:2 primarily:1 opening:1 individual:1 fitness:14 attractor:1 acceptance:1 mixture:1 tj:1 closer:2 shorter:1 unless:1 old:2 desired:1 eckhard:1 increased:1 werner:1 lattice:2 neutral:1 uniform:2 johnson:1 varies:1 international:1 sensitivity:4 probabilistic:2 together:1 central:1 containing:1 possibly:1 worse:2 corner:10 inefficient:2 til:1 li:1 diversity:1 satisfy:3 depends:2 performed:1 root:3 break:2 view:1 reached:1 start:1 parallel:1 xlx:4 mutation:13 square:3 il:1 ni:6 variance:3 largely:1 characteristic:1 yield:2 spaced:1 landscape:2 generalize:2 produced:1 none:1 eda:5 ed:1 against:1 frequency:1 bioscience:1 sampled:1 fiat:2 actually:1 back:1 appears:1 response:1 improved:4 just:1 until:4 hand:2 interfere:1 perhaps:1 behaved:1 building:1 effect:2 contain:2 true:1 requiring:1 former:1 evolution:1 lozano:1 laboratory:1 attractive:1 during:1 self:1 criterion:3 consideration:2 exponentially:5 jl:1 analog:1 kluwer:1 pell:1 mellon:1 haystack:10 tuning:2 illinois:2 stable:1 etc:1 add:1 posterior:1 recent:1 wellknown:1 inequality:1 binary:2 yi:10 greater:5 recognized:1 converge:3 ii:8 multiple:1 smooth:3 technical:2 academic:1 long:2 lin:1 equally:2 visit:1 halt:1 hasting:1 prediction:3 expectation:3 ion:1 proposal:1 whereas:1 grow:1 source:2 publisher:1 biased:2 finely:1 unlike:1 fell:3 member:2 call:1 integer:1 near:3 constraining:1 enough:4 fit:1 repeatedly:2 pelikan:1 ignored:1 dramatically:1 detailed:16 clear:1 amount:3 simplest:1 generate:2 shapiro:3 reinforces:1 carnegie:1 affected:1 putting:1 key:1 drawn:2 changing:1 rrf:1 graph:2 fraction:1 sum:1 run:8 inverse:1 parameterized:1 compete:1 counteract:1 j2u:1 almost:2 reasonable:1 scaling:6 entirely:1 fl:1 flat:7 markus:1 generates:1 argument:1 extremely:1 min:2 expanded:2 department:2 remain:1 smaller:2 increasingly:1 ascertain:1 across:1 metropolis:1 making:1 happens:1 projecting:1 invariant:1 equation:12 previously:2 fail:1 mechanism:1 obey:3 away:2 appropriate:3 enforce:1 distinguished:1 robustness:1 hat:1 denotes:2 binomial:4 graphical:4 jlog:1 approximating:1 hypercube:7 unchanged:1 move:5 quantity:2 occurs:2 dependence:1 usual:1 evolutionary:2 win:1 subspace:2 distance:2 unstable:3 assuming:1 relationship:1 balance:16 difficult:3 i_:1 av:1 markov:1 finite:1 gas:3 situation:1 variability:1 arbitrary:1 drift:8 introduced:1 required:1 learned:1 usually:1 below:3 including:1 event:1 difficulty:1 treated:1 recursion:2 mn:1 scheme:1 improve:1 altered:1 misleading:1 deviate:1 prior:3 review:1 understanding:1 loss:1 lecture:1 generation:1 interesting:1 analogy:1 versus:2 consistent:1 lo:2 genetics:1 repeat:2 lln:1 cmucs:1 bias:2 exponentiated:1 curve:1 dimension:1 transition:1 avl:1 lnl:1 made:2 clue:1 confirm:2 global:7 reveals:1 a_:1 decrementing:1 xi:5 search:9 continuous:1 learn:2 robust:1 m13:1 complex:2 did:3 motivation:1 big:1 revisits:1 incrementing:1 allowed:1 site:2 fails:1 candidate:1 lie:2 vanish:4 removing:2 down:2 formula:1 decay:1 evidence:1 importance:1 easier:1 rejection:2 entropy:1 likely:1 failed:1 monotonic:1 corresponds:1 towards:14 man:1 bennett:1 change:3 hard:1 baluja:2 except:1 determined:1 unpredictably:1 averaging:1 called:2 accepted:1 tendency:1 latter:1 jonathan:1 phenomenon:2 ex:1 |
1,250 | 2,139 | Stability-Based Model Selection
Tilman Lange, Mikio L. Braun, Volker Roth, Joachim M. Buhmann
(lange,braunm,roth,jb)@cs.uni-bonn.de
Institute of Computer Science, Dept. III,
University of Bonn
R?omerstra?e 164, 53117 Bonn, Germany
Abstract
Model selection is linked to model assessment, which is the problem of
comparing different models, or model parameters, for a specific learning
task. For supervised learning, the standard practical technique is crossvalidation, which is not applicable for semi-supervised and unsupervised
settings. In this paper, a new model assessment scheme is introduced
which is based on a notion of stability. The stability measure yields an
upper bound to cross-validation in the supervised case, but extends to
semi-supervised and unsupervised problems. In the experimental part,
the performance of the stability measure is studied for model order selection in comparison to standard techniques in this area.
1 Introduction
One of the fundamental problems of learning theory is model assessment: Given a specific
data set, how can one practically measure the generalization performance of a model trained
to the data. In supervised learning, the standard technique is cross-validation. It consists in
using only a subset of the data for training, and then testing on the remaining data in order to
estimate the expected risk of the predictor. For semi-supervised and unsupervised learning,
there exist no standard techniques for estimating the generalization of an algorithm, since
there is no expected risk. Furthermore, in unsupervised learning, the problem of model
order selection arises, i.e. estimating the ?correct? number of clusters. This number is part
of the input data for supervised and semi-supervised problems, but it is not available for
unsupervised problems.
We present a common point of view, which provides a unified framework for model assessment in these seemingly unrelated areas of machine learning. The main idea is that an
algorithm generalizes well, if the solution on one data set has small disagreement with the
solution on another data set. This idea is independent of the amount of label information
which is supplied to the problem, and the challenge is to define disagreement in a meaningful way, without relying on additional assumptions, e.g. mixture densities. The main
emphasis lies on developing model assessment procedures for semi-supervised and unsupervised clustering, because a definitive answer to the question of model assessment has
not been given in these areas.
In section 3, we derive a stability measure for solutions to learning problems, which allows us to characterize the generalization in terms of the stability of solutions on different
sets. For supervised learning, this stability measure is an upper bound to the 2-fold cross-
validation error, and can thus be understood as a natural extension of cross-validation to
semi-supervised and unsupervised problems.
For experiments (section 4), we have chosen the model order selection problem in the
unsupervised setting, which is one of the relevant areas of application as argued above. We
compare the stability measure to other techniques from the literature.
2 Related Work
For supervised learning problems, several notions of stability have been introduced ([10],
[3]). The focus of these works lies on deriving theoretical generalization bounds for supervised learning. In contrast, this work aims at developing practical procedures for model
assessment, which are also applicable in semi- and unsupervised settings. Furthermore, the
definition of stability developed in this paper does not build upon the cited works.
Several procedures have been proposed for inferring the number of clusters of which we
name a few here. Tibshirani et al. [14] propose the Gap Statistic that is applicable to Euclidian data only. Given a clustering solution, the total sum of within-cluster dissimilarities is
computed. This quantity computed on the original data is compared with the average over
data which was uniformly sampled from a hyper-rectangle containing the original data.
The which maximizes the gap between these two quantities is the estimated number of
clusters. Recently, resampling-based approaches for model order selection have been proposed that perform model assessment in the spirit of cross validation. These approaches
share the idea of prediction strength or replicability as a common trait. The methods exploit the idea that a clustering solution can be used to construct a predictor, in order to
compute a solution for a second data set and to compare the computed and predicted class
memberships for the second data set. In an early study, Breckenridge [4] investigated the
usefulness of this approach (called replication analysis there) for the purpose of cluster
validation. Although his work does not lead to a directly applicable procedure, in particular
not for model order selection, his study suggests the usefulness of such an approach for
the purpose of validation. Our method can be considered as a refinement of his approach.
Fridlyand and Dudoit [6] propose a model order selection procedure, called Clest, that also
builds upon Breckenridge?s work. Their method employs the replication analysis idea by
repeatedly splitting the available data into two parts. Free parameters of their method are
the predictor, the measure of agreement between a computed and a predicted solution and
a baseline distribution similar to the Gap Statistic. Because these three parameters largely
influence the assessment, we consider their proposal more as a conceptual framework than
as a concrete model order estimation procedure. In particular, the predictor can be chosen
independent of the clustering algorithm which can lead to unreliable results (see section
3). For the experiments in section 4, we used a linear discriminant analysis classifier, the
Fowlkes-Mellows index for solution comparison (c.f. [9, 6]) and the baseline distribution of
the Gap Statistic. Tibshirani et al. [13] formulated a similar method (Prediction Strength)
for inferring the number of clusters which is based on using nearest centroid predictors.
Roughly, their measure of agreement quantifies the similarity of two clusters in the computed and in the predicted solution. For inferring a number of clusters, the least similar
pair of clusters is taken into consideration. The estimated is the largest for which the
similarity is above some threshold value. Note that the similarity for
is always above
this threshold.
3 The Stability Measure
We begin by introducing a stability measure for supervised learning. Then, the stability
measure is generalized to semi-supervised and unsupervised settings. Necessary modifications for model order selection are discussed. Finally, a scheme for practical estimation of
the stability is proposed.
&#(' )
*
! " #%$ +-, $/.
' )0 *
1 + , 32546 + 7 98 9: ! "
4
4<898>=? A@BC8ED 8F=HG 8IJ
D 8F= ;
;
8 8= 8F= =
@ B 8KLD 8F= = GM @ B 8ND 8F= GPO @ B 8Q=ND 8F= = G
8ND 8F= =
8=LD 8 8F=LD 8F= =
=
+FR
+R
+ RFS K = +
1PR S +)R TKUWV @ B +)R = 6ED = GXM 1PR S +)R S O T UWV @ B +)R = 6 D +)R S = G (1)
We call the second term the stability of the predictor + and denote its expectation by Y + :
(2)
Y + , KZ\[ T UWV @ B +)R = ] D +)R S = GC^
; means perfect stability
We call the value of Y + stability cost to stress the fact that Y
and
of Y mean
with respect to and
= onlargebothvalues
Z large 6`ainstability.
Z
1PR +)R Taking
9 M ZXY expectations
+ . If + 9 iscobtained
sides yields 1 +_R
risk
b , then 1&R +)R ZXd?efgih)byj 1Pempirical
R + 9 M
dWeFfgih)minimization
jkZ
1PR + 9 over
Ed?eFfsome
gih)j 1 hypothesis
+ , and onesetobtains
Z 1 + R l`mgidWeFh)fj 1 + M Z 1 + R l`nZ
1 R + R 9 M Y + 0
(3)
Stability and Supervised Learning The supervised learning problem is defined as follows. Let
be a sequence of random variables where
are drawn i.i.d. from some probability distribution
. The
are the objects and
are the labels. The task is to find a labeling function
which minimizes the expected risk, given by
, using
only a finite sample of data as input. Here is the so-called loss function. For classification, we take the - -loss defined by
iff
and
else.
A measure of the stability of the labeling function learned is derived as follows. Note that
, since
for three labels , and , it holds that
implies
or
. Now let and be two data sets drawn independently
from the same source, and denote the predictor trained on by . Then, the test risk of
can be bounded by introducing
:
By eq. (3), the stability defined in (2) yields an upper bound on the generalization error.
It can be shown that there exists a converse upper bound, if the minimum is unique and
well-separated, such that
implies
.
Note that the stability measures the disagreement between labels on training data and test
data, both assigned by . This asymmetry arises naturally and directly measures the generalization performance of . Furthermore, the stability can be interpreted as the expected
empirical risk of with respects to the labels computed by itself (compare (1) and (2)).
Therefore, stability measures the self-consistency of . This interpretation is also valid in
the semi-supervised and unsupervised settings. Practical evaluation of the stability amounts
to 2-fold cross-validation. No improvement can therefore be expected in this area. However,
unlike cross-validation, stability can also be defined in settings where no label information
is available. This property of the method will be discussed in the remainder of this section.
Z 1 _+ R o.JdWeFfgh)j 1 +
+
+ +
Y . ;
+
Semi-supervised Learning Semi-supervised learning problems are defined as follows.
of an object
might not be known. This fact is encoded by setting
The label
, since is not a valid label. At least one labeled point must be given for every class.
Furthermore, for the present discussion, we assume that we do not have a fully labeled data
set for testing purposes.
There exist two alternatives in defining the solution to a semi-supervised learning problem.
In the first alternative, the solution is a labeling function defined on the whole object
space as in supervised learning. Then, the stability (eq. (2)) can be readily computed and
measures the confidence for the (unknown) training error.
The second alternative is that the solution is not given by a labeling function on the whole
object space, but only by a labeling function on the training set . Labeling functions
;
$
;
+
R
which are defined on the training set only will be denoted by to stress the difference. The
labeling on will be denoted by
, which is only defined on . As mentioned above,
the stability compares labels given to the training data with predicted labels. In the current
setting, there are no predicted labels, because is defined on the
training set only. One
possibility to obtain predicted labels is to introduce a predictor , which is trained using
to predict labels on the new set . Leaving as a free parameter, we define the
stability for semi-supervised learning as
R
@ B
K
Z
[
Y
, T UW V
semi
=
= ] D RFS = GC^F
(4)
Of course, the choice of influences the value of the stability. We need a condition
on the
prediction step to select . First note that (4) is the expected empirical risk of with respect
to the data source
. Analogously to supervised learning, the minimal attainable
stability
measures the extent to which classes overlap, or how consistent
s
emi
the labels are. Therefore, should be chosen to minimize s emi . Unfortunately, the
construction of non-asymptotically Bayes optimal learning algorithms is extremely difficult
and, therefore, we should not expect that
there exists a universally applicable constructive
procedure for automatically building given an .
In practice, some has to be chosen. This choice will yield larger stability costs, i.e. worse
stability, and can therefore not fake stability. Furthermore, it is often possible to construct
good predictors in practice.
Note that (4) measures the mismatch between the label genera
tor
and
the
predictor
.
Intuitively,
can lead to good stability only if the strategy of and
are similar. For unsupervised learning, as discussed in the next paragraph, the choices for
various standard techniques are natural. For example, -means clustering suggests to use
nearest centroid classification. Minimum spanning tree type clustering algorithms suggest
nearest neighbor classifiers, and finally, clustering algorithms which fit a parametric density
model should use the class posteriors computed by the Bayes rule for prediction.
dWe Y K lR
Y
E
0
Unsupervised Learning The unsupervised learning setting
is given as the problem of
. The solution
is again a function only defined
labeling a finite data set
on . From this definition, it becomes clear that we again need a predictor as in the second
alternative of semi-supervised learning.
For unsupervised learning, another problem arises. Since no specific label values are prescribed for the classes, label indices might be permuted from one instance to another, even
when the partitioning is identical. For example, keeping the same classes, exchanging the
class labels and leads to a new partitioning, which is not structurally different. In other
words, label values are only known up to a permutation. In view of this non-uniqueness of
the representation of a partitioning, we define the permutation relating indices on the first
set to the second set by the one which maximizes the agreement between the classes. The
stability then reads
@B = ] D = G ^F
Z
?
d
e
[
Y ,
h T UWV
(5)
S
Note that the minimization
take place inside the expectation, becausethe permutation
K = .hasIn topractice,
depends on the data
it is not necessary to compute
all permutations,
un
because the problem is solvable by the Hungarian method in
[11].
Model Order Selection The problem of model order selection consists in determining
the number of clusters to be estimated, and exists only in unsupervised learning.
The range of the stability depends on , therefore stability values cannot be compared
for different values of . For unsupervised
learning, the stability minimized over is
bounded from above by
, since for a larger instability, there exists a relabeling
Y
`
Y
which has smaller stability costs. This stability value is asymptotically achieved by the
random predictor which assigns uniformly drawn labels to objects. Normalizing by
the stability of the random predictor yields values independent of . We thus define the
re-normalized stability as
(6)
un
un
un
Y Y Y
9
Resampling Estimate of the Stability In practice, a finite data set
is given,
and the best model should be estimated. The stability is defined in terms of an expectation,
which has to be estimated for practical applications. Estimation of over a hypothesis set
is feasible if has finite VC-dimension, since the VC-dimension for estimating is the
same as for the empirical risk, a fact which is not proved here. In order to estimate the
stability, we propose the following resampling scheme: Iteratively split the data set into
disjoint halves, and compare the solutions on these sets as defined above for the respective
cases. After the model having the smallest value of is determined, train this model again
on the whole data to obtain the result.
Note that it is necessary to split into disjoint subsets, because common points potentially increase the stability artificially. Furthermore, unlike in cross-validation, both sets must have
the same size, because both are used as inputs to training algorithms. For semi-supervised
and unsupervised learning, the comparison might entail predicting labels on a new set, and
for the latter also minimizing over permutation of labels.
b
Y
b
Y
Y
4 Stability for Model Order Selection in Clustering: Experimental
Results
We now provide experimental evidence for the usefulness of our approach to model order
selection, which is one of the hardest model assessment problems. First, the algorithms are
compared for toy data, in order to study the performance of the stability measure under
well-controlled conditions. However, for real-world applications, it does not suffice to be
better than competitors, but one has to provide solutions which are reasonable within the
framework of the application. Therefore, in a second experiment, the stability measure is
compared to the other methods for the problem of clustering gene expression data.
Experiments are conducted using a deterministic annealing variant of -means [12] and
Path-Based Clustering [5] optimized via an agglomerative heuristic. For all data sets, we
average over resamples for
. For the Gap Statistic and Clest1 random
samples are drawn from the baseline. For Clest and Prediction Strength, the number of
resamples
is chosen the same as for our method. The threshold for Prediction Strength is
set to
. As mentioned above, the nearest centroid classifier is employed for the purpose
of prediction when using -means, and a variant of the nearest neighbor classifier is used
for Path-Based Clustering which can be regarded as a combination of Minimum Spanning
Tree clustering and Pairwise Clustering [5, 8].
We compare the proposed stability index of section 3 with the Gap Statistic, Clest and with
Tibshirani?s Prediction Strength method using two toy data sets and a microarray data set
taken from [7]. Table 1 summarizes the estimated number of clusters of each method.
;
;
;
;
;
Toy Data Sets The first data set consists of three fairly well separated point clouds, generated from three Gaussian distributions ( points from the first and the second and
in figure 1(a),
points from the third were drawn). Note that for some , for example
the variance in the stability over different resamples is quite high. This effect is due to the
model mismatch, since for
, the clustering of the three classes depends highly on the
subset selected in the resampling. This means that besides the absolute value of the stability
1
See section 2 for a brief overview over these techniques.
Data Set
Stability
Method
3 Gaussians
3 Rings -means
3 Rings Path-Based
Golub et al. data
Gap
Statistic
;
Clest
Prediction
Strength
?true?
number
or
Table 1: The estimated model orders for the two toy and the microarray data set.
costs, additional information about the fit can be obtained from the distribution of the stability costs over the resampled subsets. For this data set, all methods under comparison are
able to infer the ?true? number of clusters
. Figures 1(d) and 1(a) show the clustered
data set and the proposed stability index. For
, the stability is relatively high, which
is due to the hierarchical structure of the data set, which enables stable merging of the two
smaller sub-clusters.
In the ring data set (depicted in figures 1(e) and 1(f)), one can naturally distinguish three
ring shaped clusters that violate the modeling assumptions of -means since clusters are not
spherically distributed. Here, -means is able to identify the inner circle as a cluster with
. Thus, the stability for this number of clusters is highest (figure 1(b)). All other
methods except Clest infer
for this data set with -means. Applying the proposed
stability estimator with Path-Based Clustering on the same data set yields highest stability
for
, the ?correct? number of clusters (figures 1(f) and 1(c)). Here, all other methods
. The Gap Statistic fails here because it directly incorporates the
fail and estimate
assumption of spherically distributed data. Similarly, the Prediction Strength measure and
Clest (in the form we use here) use classifiers that only support linear decision boundaries
which obviously cannot discriminate between the three ring-shaped clusters. In all these
cases, the basic requirement for a validation scheme is violated, namely that it must not
incorporate additional assumptions about the group structure in a data set that go beyond the
ones of the clustering principle employed. Apart from that, it is noteworthy that the stability
with -means is significantly worse than the one achieved with Path-Based Clustering,
which indicates that the latter is the better choice for this data set.
Application to Microarray Data Recently, several authors have investigated the possibility of identifying novel tumor classes based solely on gene expression data [7, 2, 1].
Golub et al. [7] studied in their analysis the problem of classifying and clustering acute
leukemias. The important question of inferring an appropriate model order remains unaddressed in their article and prior knowledge is used instead. In practice however, such
knowledge is often not available.
Acute leukemias can be roughly divided into two groups, acute myeloid leukemia (AML)
and acute lymphoblastic leukemia (ALL) where the latter can furthermore be subdivided
into B-cell ALL and T-cell ALL. Golub et al. used a data set of 72 leukemia samples (25
AML, 47 ALL of which 38 are B-cell ALL samples)2 . For each sample, gene expression
was monitored using Affymetrix expression arrays.
We apply the preprocessing steps as in Golub et al. resulting in a data set consisting of 3571
genes and 72 samples. For the purpose of cluster analysis, the feature set was additionally
reduced by only retaining the 100 genes with highest variance across samples. This step
is adopted from [6]. The final data set consists of 100 genes and 72 samples. We have
performed cluster analysis using -means and the nearest centroid rule. Figure 2 shows
2
Available at http://www-genome.wi.mit.edu/cancer/
0.8
1
0.7
0.5
0.9
0.8
0.6
0.4
0.7
0.5
0.4
index
index
index
0.6
0.5
0.4
0.3
0.3
0.2
0.3
0.2
0.2
0.1
0.1
0.1
0
0
2
3
4
5
6
7
number of clusters
8
9
0
2
10
(a) The stability index for
the Gaussians data set
with -means.
3
4
5
6
7
number of clusters
8
9
10
2
(b) The stability index for
the three-ring data set with
-means Clustering.
4
5
6
7
number of clusters
8
9
10
(c) The stability index for
the three-ring data set with
Path-Based Clustering.
8
3
5
5
4
4
3
3
2
2
1
1
6
4
2
0
0
?1
?1
?2
?2
?3
?3
0
?2
?4
?4
?6
?4
?2
0
2
4
6
(d) Clustering solution on
the full data set for
.
8
?5
?5
?4
?4
?3
?2
?1
0
1
2
3
4
(e) Clustering solution on
the full data set for
.
5
?5
?5
?4
?3
?2
?1
0
1
2
3
4
5
(f) Clustering solution on
the full data set for
.
Figure 1: Results of the stability index on the toy data (see section 4).
, we estimate the highest stability. We expect
the corresponding stability curve. For
separates AML, B-cell ALL
ALL samples from each
that clustering with
and
ofT-cell
other. With respect to the known ground-truth labels,
the samples (66 samples) are
correctly classified (the Hungarian method is used to map the clusters to the ground-truth).
Of the competitors, only Clest is able to infer the ?correct? number of cluster
while
the Gap Statistic largely overestimates the number of clusters. The Prediction strength does
not provide any reasonable result as it estimates
. Note, that for
similar stability
is achieved. We cluster the data set again
for
and
compare
the
result
with the ALL ?
of the samples (62 samples) are correctly
AML labeling of the data. Here,
identified.
We conclude that our method is able to infer biologically relevant model orders. At the
same time, a is suggested that leads to high accuracy w.r.t. the ground-truth. Hence, our
re-analysis demonstrates that we could have recovered a biologically meaningful grouping
in a completely unsupervised manner.
C
5 Conclusion
The problem of model assessment was addressed in this paper. The goal was to derive a
common framework for practical assessment of learning models. Starting with defining a
stability measure in the context of supervised learning, this measure was generalized to
semi-supervised and unsupervised learning. The experiments concentrated on model or-
0.7
0.6
index
0.5
0.4
0.3
0.2
0.1
0
2
3
4
5
6
7
number of clusters
8
9
10
Figure 2: Resampled stability for the leukemia dataset vs. number of classes (see sec. 4).
der selection for unsupervised learning, because this is the area where the need for widely
applicable model assessment strategies is highest. On toy data, the stability measure outperforms other techniques, when their respective modeling assumptions are violated. On
real-world data, the stability measure compares favorably to the best of the competitors.
Acknowledgments. This work has been supported by the German Research Foundation
(DFG), grants #Buh 914/4, #Buh 914/5.
References
[1] A. A. Alizadeh et al. Distinct types of diffuse large b-cell lymphoma identified by gene expression profiling. Nature, 403:503 ? 511, 2000.
[2] M. Bittner et al. Molecular classification of cutaneous malignant melanoma by gene expression
profiling. Nature, 406(3):536 ? 540, 2000.
[3] O. Bousquet and A. Elisseeff. Stability and generalization. Journal of Machine Learning
Research, 2:499?526, 2002.
[4] J. Breckenridge. Replicating cluster analysis: Method, consistency and validity. Multivariate
Behavioural research, 1989.
[5] B. Fischer, T. Z?oller, and J. M. Buhmann. Path based pairwise data clustering with application to
texture segmentation. In LNCS Energy Minimization Methods in Computer Vision and Pattern
Recognition. Springer Verlag, 2001.
[6] J. Fridlyand and S. Dudoit. Applications of resampling methods to estimate the number of
clusters and to improve the accuracy of a clustering method. Technical Report 600, Statistics
Department, UC Berkeley, September 2001.
[7] T.R. Golub et al. Molecular classification of cancer: Class discovery and class prediction by
gene expression monitoring. Science, pages 531 ? 537, October 1999.
[8] T. Hofmann and J. M. Buhmann. Pairwise data clustering by deterministic annealing. IEEE
PAMI, 19(1), January 1997.
[9] A. K. Jain and R. C. Dubes. Algorithms for Clustering Data. Prentice-Hall, Inc., 1988.
[10] Michael J. Kearns and Dana Ron. Algorithmic stability and sanity-check bounds for leave-oneout cross-validation. In Computational Learing Theory, pages 152?162, 1997.
[11] H.W. Kuhn. The hungarian method for the assignment problem. Naval Res. Logist. Quart.,
2:83?97, 1955.
[12] K. Rose, E. Gurewitz, and G. C. Fox. A deterministic annealing approach to clustering. Pattern
Recognition Letters, 11(9):589 ? 594, 1990.
[13] R. Tibshirani, G. Walther, D. Botstein, and P. Brown. Cluster validation by prediction strength.
Technical report, Statistics Department, Stanford University, September 2001.
[14] R. Tibshirani, G. Walther, and T. Hastie. Estimating the number of clusters via the gap statistic.
Technical report, Statistics Department, Stanford University, March 2000.
| 2139 |@word nd:3 elisseeff:1 attainable:1 myeloid:1 euclidian:1 ld:3 affymetrix:1 outperforms:1 current:1 comparing:1 recovered:1 must:3 readily:1 hofmann:1 enables:1 resampling:5 v:1 half:1 selected:1 lr:1 provides:1 ron:1 replication:2 learing:1 consists:4 walther:2 paragraph:1 inside:1 manner:1 introduce:1 pairwise:3 expected:6 roughly:2 relying:1 automatically:1 becomes:1 begin:1 estimating:4 unrelated:1 bounded:2 maximizes:2 suffice:1 interpreted:1 minimizes:1 developed:1 unified:1 berkeley:1 every:1 braun:1 classifier:5 demonstrates:1 partitioning:3 converse:1 grant:1 overestimate:1 understood:1 melanoma:1 path:7 solely:1 noteworthy:1 pami:1 might:3 emphasis:1 nz:1 studied:2 suggests:2 genus:1 range:1 practical:6 unique:1 acknowledgment:1 testing:2 practice:4 procedure:7 lncs:1 area:6 empirical:3 significantly:1 confidence:1 word:1 suggest:1 cannot:2 selection:14 prentice:1 risk:8 influence:2 instability:1 applying:1 context:1 www:1 deterministic:3 map:1 roth:2 go:1 starting:1 independently:1 splitting:1 assigns:1 identifying:1 rule:2 estimator:1 array:1 regarded:1 deriving:1 his:3 stability:73 notion:2 construction:1 hypothesis:2 agreement:3 recognition:2 labeled:2 cloud:1 highest:5 mentioned:2 rose:1 trained:3 upon:2 completely:1 various:1 train:1 separated:2 distinct:1 jain:1 labeling:9 hyper:1 lymphoma:1 quite:1 encoded:1 larger:2 heuristic:1 widely:1 sanity:1 stanford:2 statistic:12 fischer:1 pempirical:1 itself:1 final:1 seemingly:1 obviously:1 sequence:1 propose:3 fr:1 remainder:1 relevant:2 iff:1 crossvalidation:1 cluster:33 asymmetry:1 requirement:1 perfect:1 ring:7 leave:1 object:5 derive:2 dubes:1 nearest:6 eq:2 c:1 predicted:6 implies:2 hungarian:3 kuhn:1 correct:3 aml:4 vc:2 lymphoblastic:1 eff:1 byj:1 argued:1 subdivided:1 generalization:7 clustered:1 extension:1 hold:1 practically:1 considered:1 ground:3 hall:1 algorithmic:1 predict:1 tor:1 early:1 smallest:1 purpose:5 uniqueness:1 estimation:3 applicable:6 label:23 largest:1 minimization:3 mit:1 always:1 gaussian:1 aim:1 volker:1 derived:1 focus:1 joachim:1 improvement:1 naval:1 indicates:1 check:1 contrast:1 centroid:4 baseline:3 membership:1 germany:1 classification:4 denoted:2 retaining:1 fairly:1 uc:1 construct:2 having:1 shaped:2 identical:1 hardest:1 unsupervised:21 leukemia:6 minimized:1 jb:1 report:3 few:1 employ:1 relabeling:1 gpo:1 dfg:1 consisting:1 possibility:2 highly:1 evaluation:1 golub:5 mixture:1 hg:1 necessary:3 respective:2 fox:1 tree:2 re:3 circle:1 theoretical:1 minimal:1 instance:1 modeling:2 exchanging:1 assignment:1 cost:5 introducing:2 subset:4 predictor:13 usefulness:3 conducted:1 characterize:1 answer:1 density:2 fundamental:1 cited:1 michael:1 analogously:1 concrete:1 again:4 containing:1 worse:2 toy:6 de:1 sec:1 inc:1 depends:3 performed:1 view:2 linked:1 bayes:2 minimize:1 accuracy:2 variance:2 largely:2 yield:6 identify:1 monitoring:1 zx:1 classified:1 ed:3 definition:2 competitor:3 energy:1 naturally:2 monitored:1 sampled:1 proved:1 dataset:1 knowledge:2 segmentation:1 supervised:28 botstein:1 furthermore:7 assessment:13 replicability:1 building:1 effect:1 validity:1 normalized:1 true:2 name:1 brown:1 hence:1 assigned:1 read:1 spherically:2 iteratively:1 self:1 generalized:2 stress:2 fj:1 resamples:3 consideration:1 novel:1 recently:2 fridlyand:2 common:4 permuted:1 overview:1 discussed:3 interpretation:1 relating:1 trait:1 breckenridge:3 consistency:2 similarly:1 replicating:1 stable:1 entail:1 similarity:3 acute:4 posterior:1 multivariate:1 apart:1 logist:1 verlag:1 wv:4 der:1 minimum:3 additional:3 employed:2 semi:17 violate:1 full:3 infer:4 technical:3 constructive:1 profiling:2 cross:9 divided:1 molecular:2 controlled:1 prediction:13 variant:2 basic:1 vision:1 expectation:4 achieved:3 cell:6 proposal:1 tku:1 annealing:3 addressed:1 else:1 source:2 leaving:1 microarray:3 unlike:2 incorporates:1 spirit:1 unaddressed:1 call:2 iii:1 split:2 fit:2 hastie:1 identified:2 lange:2 idea:5 inner:1 expression:7 repeatedly:1 fake:1 clear:1 quart:1 amount:2 gih:3 concentrated:1 reduced:1 http:1 supplied:1 exist:2 estimated:7 disjoint:2 tibshirani:5 correctly:2 group:2 threshold:3 drawn:5 rectangle:1 uw:1 asymptotically:2 sum:1 letter:1 extends:1 place:1 reasonable:2 decision:1 summarizes:1 bound:6 resampled:2 distinguish:1 fold:2 strength:9 diffuse:1 bousquet:1 bonn:3 emi:2 extremely:1 prescribed:1 buh:2 relatively:1 department:3 developing:2 combination:1 march:1 smaller:2 across:1 wi:1 oller:1 modification:1 biologically:2 cutaneous:1 intuitively:1 pr:4 taken:2 behavioural:1 remains:1 german:1 fail:1 malignant:1 adopted:1 available:5 generalizes:1 gaussians:2 apply:1 hierarchical:1 appropriate:1 disagreement:3 fowlkes:1 alternative:4 original:2 remaining:1 clustering:29 exploit:1 build:2 question:2 quantity:2 strategy:2 parametric:1 dwe:1 september:2 separate:1 agglomerative:1 extent:1 discriminant:1 spanning:2 besides:1 index:13 minimizing:1 difficult:1 unfortunately:1 october:1 potentially:1 favorably:1 oneout:1 unknown:1 perform:1 upper:4 finite:4 january:1 tilman:1 defining:2 gc:2 introduced:2 pair:1 namely:1 optimized:1 learned:1 able:4 beyond:1 suggested:1 pattern:2 mismatch:2 oft:1 challenge:1 rf:2 overlap:1 natural:2 predicting:1 buhmann:3 solvable:1 scheme:4 improve:1 brief:1 gurewitz:1 prior:1 literature:1 discovery:1 determining:1 loss:2 fully:1 expect:2 permutation:5 dana:1 validation:13 foundation:1 consistent:1 article:1 principle:1 classifying:1 share:1 cancer:2 course:1 supported:1 free:2 keeping:1 side:1 institute:1 neighbor:2 taking:1 absolute:1 distributed:2 boundary:1 dimension:2 curve:1 valid:2 world:2 genome:1 kz:1 author:1 refinement:1 universally:1 preprocessing:1 uni:1 unreliable:1 gene:9 conceptual:1 conclude:1 un:4 quantifies:1 table:2 additionally:1 nature:2 investigated:2 artificially:1 main:2 whole:3 definitive:1 mikio:1 alizadeh:1 structurally:1 inferring:4 sub:1 fails:1 lie:2 third:1 specific:3 normalizing:1 evidence:1 exists:4 grouping:1 merging:1 texture:1 dissimilarity:1 gap:10 depicted:1 springer:1 truth:3 goal:1 formulated:1 dudoit:2 feasible:1 determined:1 except:1 uniformly:2 tumor:1 kearns:1 total:1 called:3 discriminate:1 experimental:3 meaningful:2 select:1 support:1 latter:3 arises:3 violated:2 incorporate:1 dept:1 |
1,251 | 214 | A Neural Network to Detect Homologies in Proteins
A Neural Network to Detect
Homologies in Proteins
Yoshua Bengio
Yannick Pouliot
School of Computer Science
McGill University
Montreal, Canada H3A 2A7
Department of Biology
McGill University
Montreal Neurological Institute
Samy Bengio
Patrick Agin
Departement dlnformatique
Universite de Montreal
Departement d'Informatique
U niversite de Montreal
.A BSTRACT
In order to detect the presence and location of immunoglobulin (Ig) domains from amino acid sequences we built a system
based on a neural network with one hidden layer trained with
back propagation. The program was designed to efficiently
identify proteins exhibiting such domains, characterized by a
few localized conserved regions and a low overall homology.
When the National Biomedical Research Foundation (NBRF)
NEW protein sequence database was scanned to evaluate the
program's performance, we obtained very low rates of false
negatives coupled with a moderate rate of false positives.
1 INTRODUCTION
Two amino acid sequences from proteins are homologous if they can be
aligned so that many corresponding amino acids are identical or have similar
chemical properties. Such subsequences (domains) often exhibit similar three
dimensional structure. Furthemore, sequence similarity often results from
common ancestors. Immunoglobulin (Ig) domains are sets of ,a-sheets bound
423
424
Bengio, Bengio, Pouliot and Agin
by cysteine bonds and with a characteristic tertiary structure. Such domains
are found in many proteins involved in immune, cell adhesion and receptor
functions. These proteins collectively form the immunoglobulin superfamily
(for review, see Williams and Barclay, 1987). Members of the superfamily
often possess several Ig domains. These domains are characterized by wellconserved groups of amino acids localized to specific subregions. Other residues outside of these regions are often poorly conserved, such that there is
low overall homology between Ig domains, even though they are clearly
members of the same superfamily.
Current search programs incorporating algorithms such as the Wilbur-Lipman
algorithm (1983) or the Needleman-Wunsch algorithm (1970) and its modification by Smith and Waterman (1981) are ill-designed for detecting such
domains because they implicitly consider each amino acid to be equally important. This is not the case for residues within domains such as the Ig
domain, since only some amino acids are well conserved, while most are variable. One solution to this problem are search algorithms based upon the statistical occurrence of a residue at a particular position (Wang et al., 1989;
Gribskov et al., 1987). The Profile Analysis set of programs published by the
University of Wisconsin Genetics Computer Group (Devereux et al., 1984)
rely upon such an algorithm. Although Profile Analysis can be applied to
search for domains (c./. Blaschuk, Pouliot & Holland 1990), the output from
these programs often suffers from a high rate of false negatives and positives.
Variations in domain length are handled using the traditional method of penalties proportional to the nuinber of gaps introduced, their length and their position. This approach entails a significant amount of spurious recognition if
there is considerable variation in domain length to be accounted for.
We have chosen to address these problems by training a neural network to
recognize accepted Ig domains. Perceptrons and various types of neural networks have been used previously in biological research with various degrees of
success (cf. Stormo et al., 1982; Qian and Sejnowski, 1988). Our results suggest that they are well suited for detecting relatively cryptic sequence patterns
such as those which characterize Ig domains. Because the design and training
procedure described below is relatively simple, network-based search programs constitute a valid solution to problems such as searching for proteins
assembled from the duplication of a domain.
2 ALGORITHM, NETWORK DESIGN AND TRAINING
The network capitalizes upon data concerning the existence and localization
of highly conserved groups of amino acids characteristic of the Ig domain. Its
design is similar in several respects to neural networks we have used in the
study of speech recognition (Bengio et al., 1989). Four conserved subregions
(designated P1-P4) of the Ig domain homology were identified. These roughly
correspond to ,a-strands B, C, E and F, respectively, of the Ig domain (see
also Williams and Barclay, 1988). Amino acids in these four groups are not
necessarily all conserved, but for each subregion they show a distribution very
different from the distribution generally observed elsewhere in these proteins.
Hence the first and most important stage of the system learns about these
joint distributions. The program scans proteins using a window of 5 residues.
A Neural Network to Detect Homologies in Proteins
The first stage of the system consists of a 2-layer feedforward neural network
(5 X 20 inputs - 8 hidden - 4 outputs; see Figure 1) trained with back propagation (Rumelhart et al., 1986). Better results were obtained for the recognition
of these conserved regions with this architecture than without hidden layer
(similar to a perceptron). The second stage evaluates, based upon the stream
of outputs generated by the first stage, whether and where a region similar to
the Ig domain has been detected. This stage currently uses a simple dynamic
programming algorithm, in which constraints about order of subregions and
distance between them are explicitly programmed. We force the recognizer to
detect a sequence of high values (above a threshold) for the four conserved
regions, in the correct order and such that the sum of the values obtained at
the four recognized regions is greater than a certain threshold. Weak penalties
are applied for violations of distance constraints between conserved subregions (e.g., distance between P1 and P2, P2 and P3, etc) based upon simple
rules derived from our analysis of Ig domains. These rules have little impact if
strong homologies are detected, such that the program easily handles the large
variation in domain size exhibited by Ig domains. It was necessary to explicitly formulate these constraints given the low number of training examples as
well as the assumption that the distance between groups is not a critical
discriminating factor. We have assumed that inter-region subsequences probably do not significantly influence discrimination.
4 output units
representing
4 features of
the Ig domain
8 hidden
units
20
possible
amino
acids
window scanning 5 consecutive residues
Figure 1: Structure of the neural network
425
426
Bengio, Bengio, Pouliot and Agin
filename : A22771.NEW
input sequence name: 19 epsilon chain C region - Human
HOMOLOGY starting at 24
VTLGCLATGYFPEPVMVTWDTGSLNGTTMTLPATTLTLSGHYAT1SLLTVSGAWAKQMFTC
P1
P2
P3
P4
Ending at 84. Score = 3.581
HOMOLOGY starting at 130
1QLLC LVSGYTPGT1NITWLEDGQVMDVD LSTASTTQEGE LASTQSE LTLSQKHWLSDRTYTC
P1
P2
P3
P4
Ending at 192. Score = 3.825
HOMOLOGY starting at 234
PTITCLVVDLAPSKGTVNLTWSRASGKPVNHSTRKEEKQRNGTLTVTSTLPVGTRDW1EGETYQC
P1
P2
P3
P4
Ending at 298. Score = 3.351
HOMOLOGY starting at 340
RTLACLIQNFMPED1SVQWLHNEVQLPDARHSTTQPRKTKGSGFFVFSRLEVTRAEWEQKDEF1C
P1
P2
P3
P4
Ending at 404. Score - 3.402
Figure 2: Sample output from a search of NEW. Ig domains
present within the constant region of an epsilon Ig chain
(NBRF file number A22771) are listed with the position of
P1-P4 (see text). The overall score for each domain is also listed.
As a training set we used a group of 30 proteins comprising bona fide Ig
domains (Williams and Barclay, 1987). In order to increase the size of the
training set, additional sequences were stochastically generated by substituting
residues which are not in critical positions of the domain. These substitutions
were designed not to affect the local distribution of residues to minimize
changes in the overall chemical character of the region.
The program was evaluated and optimized by scanning the NBRF protein databases (PROTEIN and NEW) version 19. Results presented below are based
upon searches of the NEW database (except where otherwise noted) and were
generated with a cutoff value of 3.0. Only complete sequences from vertebrates, insects (including Drosophila melanogaster) and eUkaryotic viruses
were scanned. This corresponds to 2422 sequences out of the 4718 present in
the NEW database. Trial runs with the program indicated that a cutoff threshold of between 2.7 and 3.0 eliminates the vast majority of false positives with
little effect upon the rate of false negatives. A sample output is listed in Figure 2.
3 RESULTS
When the NEW protein sequence database of NBRF was searched as
described above, 191 proteins were identified to possess at least one Ig
domain. A scan of the 4718 proteins comprising the NEW database required
an average of 20 hours of CPU time on a VAX 11/780. This is comparable to
other computationally intensive programs (e.g., Profile Analysis). When run
on a SUN 4 computer, similar searches required 1.3 hours of CPU time. This
is sufficiently fast to allow the user to alter the cutoff threshold repeatedly
when searching for proteins with low homology.
A Neural Network to Detect Homologies in Proteins
Table 1: Output from a search of the NEW protein sequence database.
Domains are sorted according to overall score.
3.0017 ClAss II hlstocompatlb. ant'fen, Hl,A-OR bec:a- I chain precursor (REM) . Hu,.,.n 3.4295 " bPPII chain V region - Mouse H 37-10
3.014& NonsJMdf'k: cross?ructtng an,..,. precursor? Human
3.429519 bppa chlln V region - Moule H37-&4
3.0161 ,..teffl-dertylld growth factor receptor precursor ? Moun
3.4295 Ig kappa chlln V regions - Moun Hn-C6 and H22f>2S
3.0164 Til class I hlscocomp.alib. ,nUgen. Til-, alpha chain ? Mouse
3.4331 T-uU rectPtOr alpha chain precursor V '~'on IP71) . Mouse
3.0164 Ta. class I hlstocomPliUb. ant.n. Tj? b _Iphll chain? Moust
3.0223 Vttronectln recept:or alph_ ,h.n precursor ' HUman
3.0226 T-CtllsurfKe gtycoprotetn ly-3 precursor ' Moun
3.0244 Klnase-,". trlnstormlng ploteln (srd (EC 2.7. 1.? ) . AVI,an urcomil VirUS
),0350 It alptt.." chlln C region - Humin
J.OJ50 It alptt.., I chain C regIOn - Human
3.0J?0 It alph..,2 chlln C region. A2m( I) lilorype ' Human
3.0-409 Gr.nulocyte-macroph.ge colonv?sUmulaUng flCcor I precursor - Moust
J.04I' HLA dass 1 hlstocomparlb . ant~en. Ilph. chain precursor' Human
3.0492 HADH-ubtquktOne ox~or.uctase (EC 1.6.S.3). chlln 5 - Fruit fly (Drosophila)
3.0501 NAIlH?.biq.lnOn??? Ido.ed.cu.. (Ee I.6.S.31. chain I ? F.... IIy (O.os.pllilal
3.0511 HLA clas ? htstocomp.Mlb. ant'9tft. DP bet. chain precursor - clone
3.0511 HLA cia, ? hlstocompatlb. ant...,. DP4 bet. chain ptecunor - HUmin
3.0SISHLA cia, ? hlstocomplltlb . ? nt.... OPW4 bet. I chain ptecursor - Human
3.0520 Class n histocompaUb. ? nt'gen. HLA-OQ beta ch.ln precursor (REM) - Human
].0561 rroteln' ryroSlne kinase (Ee 2. 7.1.1 12). lymphocyte - Moun
].0669 H-2 clas. hJstoc:ompaUb. ant'9tft, A?beca?2 chain ptecursor - Mouse
3.072] T-cell ree.,.or pnvna cham precursor'll 'eglOn (MNCI) - Mouse
3.072J T~ reeepeor glfTV1'\a cham ptecursor 'II regAon IRAeII} - Mouse
3.072J T-cell ree_or glfTV1'\a cha... ptecursor 'II 'eglon IRAe4) . Mouse
3.072J T-eeR ree_or glfTV1'\a chain ptecursor 'II region 'RAC42) . Mouse
3.072J T-c:eII ree_or glfTV1'\a chatn ptecursor 'II region (RACSo) . Mouse
3.0750 T-ctl r?_or bet. cha'" V region (C.F~ ? Mouse
J.07&01g hefty Chain V retlon ? Moule 251 .3
3.0711 T-col(
bOlA ch .. n "'eglon (SUp?T 'I . Hum...
3.0711 H?Z cia. I hi>.ocompotlb. ....Igen. Q7 olpllo ch.ln " ..c.rs" . Mo...
3.0717 "?2 class I hlsrocompatlb . ? nr'left, OS IIlpha ch.ln precursor - Mouse
3.0912 MytiIn-assoclatld gtycoptoteln 11236 long form precurso' - Rilt
] .0912 MyefIIt-.socl.rld g~oproteln 1&2]6 shon form precursor - R.t
3.09&2 MyoIIrtoesoc ... ed 91\'<.pr ???ln precu .. or. b,aln . Ra.
3.09&2 ~soc".ed "'90 glyc.pr ....n prec ...... Rat
3.0991 Closs I hls.ocompotlb. ""'VOn. BolA .Iph. ch.ln prec ????? (BLI?51 . Bovl....
J.099aOass I htstocompatlb_ antigen, loLA alpha chilln precursor (BU ?]) - Bovtne
J. I 04& H-2 clas I hlstocompatlb. IIntM1en. K?" a6pha chilln precursor? Mouse
J . I0&61g h.vy chain precursor'll regIOn - Mous. VCAM3 2
J. I 128 T-cell rcepeor .Ipha ,haln precursor V region (MO I 3~ - Mouse
J. II29 T<ell ree_or detta chain V region ION?4) . Moule
3.1192 T<ell rcepeor bet. chain precursor'll region IVAk) - Mouse
3.126S T -c:eU ree_or glfTV1'\a cham ptecursor 'II regIOn IK20) ? Human
3.1 J47 T-c:eU ree_or alph. chilln precursor V region (HAPOS) - Human
3. 1623 T-cen surface gtycoprotetn COl ptecurs.or . Human
3.1623: T-c:eI surface gtvcoprolelft COl prottln precunor . Human
3.1776 .... e-nma-3 chain C reQ1on . C]mlb) allotype Humin
3.1931 HypothetICal proc",n HQlf 2 ? C..,tome.;JaloY1rLls Istraln AD 169)
3.2041 SodIum channel prottln II Rat
3.20441g huvy chain'll re.;Jlon Afr"An cla*fd "og
3.2141 SURF- I protein' Mouu
].2207 T-cell recepc:or alpha chain pr~ Uls.or \0' 1f"910n (HAP 10~ . Human
3.2300 1et.-2-mlCroglobulin pre<",r~Or Hun-,an
] .2300 Beu-2-mlCroglobulln. modtflfd Human
3.2106 rreonancy-spt:Clflc bela I Qlyc opror~tn E prf'(u rs or Human
3.2344lgE Fc receptor Ilpha ,haln prKufSor Hurnan
3.2420 T-c:ell surflCe Qtycoprotetn C02 pte<unor Pat
3.2422 H?2 class N htuocompaub .nIl9~n I A I~OOI bf'ta cham precursor - Moun
3.25;2 HLA elms II hlstocompaub .ntlgen . op."..e aloha I cham precursor? Human
3.2552 HLA class II hlsro(ompat lb ,ntlgf'n ~ 8 .Ipha O'laln precursor' Human
3.2654 T-c,1/ surface glvcoproteln CO&' JI K (hJlln pfKur sor . Rat
3.2726 Myelin PO ptoteln? Bovtne
J.21141t .Iptt.., 1 chain e regton Huma"
3.21141g .Iph.., I chaIR C recJlon HumM
3.2120 Thy-I membrane glycoprotein p,e<u,~or Mouse
3.2&40 5mh clas II hlstocompilub anogen prlKurSor Ehrenberg smote-rat
1.3039 X-lInkld chronk: granulomatous dlSeast plotem Human
3.3013 rregnartCy-speclflC bela I Vl lyc oprot<t:ln ( pr<t:(unor Human
) . 3013 Prt9n ....cy-speclfiC beta' I gtycoprotetn (J pr KurSOf . Humiln
3. 30 .... T-cell recepror bell chain precUlsor \I r~IOn t 16) Human
3.3251 It pnrna- I Ill) garnma? 2b fe receptor p,KU'lor Mouse
3.]414 HypodMlkll hvbr~ IQIT?cell receplOr prftuts.or \I ff:9lon (SUp?T 1~ - Humiln
3.]414. heavy chain precursor V II recJlon Human 71 2
3.]414 Ig heavy chain precursor V II reqton Human 71 ?
3.)417 Nellral celf IdhHk)n prot~ln pfftUnOr Mou se
3. 35 If Ig epsilon chatn C recJlon Human
1 35 I I Ig epsYon chain C recJlon . HUman
3.]S22 T-c:ell rectpc:or alpha chain V reqlon (80fl alpha I) Moun
3.J605 lIft.ry gtycoptoteln I . Humil"
3.3131 T-c:eII receptor garrvnil-I chain C 'ecJlon IMNGI and MNCn - Moust
3.]131 T-c:ell ,eceplor gamma I chain C IPglOn Mouu
3.3861 T?cell 9IftVTIa chain precursor V rf:'CJlon ('II j) Moun
3.4024 Ie ep51"n chilln C recJlon . Human
3.4024" epSolkJn chain C region ? Human
3.4110 Ig heavy chain V region ? Mouse Hl6-2
3.41 3J I, heavy chlln 'II region ? Mouse H]7 ?60
3.41521g heavy chain V rec)lon . Mouse H 18-S.1 S
3.41 S5 191 kappe chlln V region? Mouse HP9
3.4171191 heavy chain'll region? Mouse If6
3.4191 Ig kappa! chain'll region ? Mouse HieS .4l) I
3.4199lg heavy cha"l V region ? Mouse ]010
3.4199 ? heavy cha," V regIon? Mouse II CR kt I I
3.4211 191 heavy Chal" V r<t:9lon ? MOllse HPll and HP27
3.421] Prt9nancy,spt:Clflc b<t:ta? I glycoprotein ( prKursor . Human
3.4213 Prt9nancy?speclfK btla? 1 g tyC OPfO(tln 0 prKursor . HUman
3421 I T?celt receplor beta chain prKUls o r V ffglon (4 C3) . Mouse
].4211 T-cell receptor beta chain precursor 1/ region (810) Mouse
34212 Sodium channel prott'ln II Rat
3 429S Ig kappa ch.ln V rt'Qlon (HZ8-A.1) Mouse H28-A2
3429519 kilppa ch-lln V r~lon . Mous.e H I S& 89H4
3.429519 kappa chain V recJlon Mouse H 37 ] I I
3.4295 Ig kappa chain V region ? MouS<t: H]] 40
3.429S Ig kappa chain V ft:qlon Mouse H 3 7 ")
) 4295 ~ k.3ppa chain V rt:910n Mouse Hll 45
,_or
34572 T?ceU surface glycoprotein CO) epsilon chain - Human
3.4594 T~en sI,.Ia,. gtycDprote .... CO. precursor? Mouse
3.4594 T'ul) surrace gtycoproteln lyt?2 precursor? Mouse
3.4595 T-c:eII recePior .. ptta chain precursor V region (HAPO$ - Humin
3.4606T-c:ell rec_or gamma-2 chlln C region eMHC& Ind MN(9) ? Mouse
3.4614 T-c:eII receptor g.nwna ch.ln C region (PfER) ? Human
3.4614 T-c:ell receJKor gamrna-I chlln C region - Hu~n
3.4614 T-cell receptOr gamrna-2 chlln C region - Human
3.4620 It heevy chain V regkln - Mouse H 146-2413
3.4620. heavy chain V region - Mouse HI 5a-I9H4
3.4620 19 heavy chain" .eglon . M???? H3S,C&
l46lO I, heavy chain Pfecursor V region? Mouse M~J3
3.4690 T-c:eII rec_or beta- I ch.ln e regIOn' Human
3.4690T-c:eIf receptor beta-I chain C regIOn? Moyse
3.4690 T-c:en receptor bK~2 chain C regIOn - Hum.n
3.4690 T-cell receptor bK~2 chain C regtOn - Human
3.4769 ? ~3 chain e reg~on. G3m(b) allOrypa - Hum.n
] .479& It k.ppa ,haln V region - Mouse H 146-2483
3.479& It k.ppa Ch.ln V region - Mouse H36-2
3.479& It kappa ch.m V region - Mouse H37-62
3.479& It kappa ch.m V region - Mouse HH?12
3.4110 It kappa chain V-I retlon . HUman WII( I)
3.48-iO Peroxklase (Ee I.Il.l.n precursor - Human
3.4&&& PIa~tv. 9rowth IKIM reeeptor precursor - Mouse
3._5 N.t<h prot..... f ?? ,. fly
3._5 N.tch pr....... f ?? I. fly
3.4983: T<" recepror beta chain precursor V rt9lon (MT I-I) - Human
3.491J T<eII receptOr beta-2 cheln precursor V regkMt MOlT' 4' Human
].4991", kappII chain Pfecursor V region - Mouse Set-.
3.S035 Alkol... pII.. pII.... (EC 3. 1.3.11 p.ecU"Of ? H.man
3.5061 "heavy choln" 'eglo.. ? M.... H 37?&2
3.SO&2 Closs R hlsllOCompotlb.
HIA-DR botaoZ ch .... proc.rs.r (REMI . H.m...
3.5012 H-2 class. hlstocomPllttb. antigen. ?.a/k bet .. 2 chain PfKyrsor - Mouse
3.5012 H-2 class nhlstoCOlnpallb. .,It~. E1I beu-2 ch.ln precursor' Mouse
] .SOI2 HLA class II hlstocompallb. anll9en, OR I beta chain (cklne 69) - Humin
3.5012 HLA class. hlsbKompKIb. antigen. OR bela chain precursor
3.S012 HLA class II hlstocompatlb_anUgtft. Ollt beta chain precursor A5) - Hum.n
3.5012 HLA class I hlstocomp.lltib. antlgen, Ollt- I bet. ch .... precursor - Human
3.5012 HLA class I hlstocompilrlb. antlten. OR-4 betll chain' Human
3.S012 HLA class h hlstocompatlb. anrlttft, DR-5 kli chain precursor' Hum.n
3.509419 IiIm~S chlln C region - Mouse
3.S 144lg .lphl?2 ch.", e region. A2m( I) alforype - Human
3.5150 Ig heavy chain V region? Mouse H2a-A2
3.5180 Biliary gtycoprDtein I- Human
3.5193 Ig heavy chain V region - Mouse H37-45
3 5193 Ig heavy chain V regions - Mouse HJ7?80 and H]7-43
35211 Ig IMnbda chain ptecursor V region' Rat
15264 Ig huvy chain V region - Mouse H ]7-62
] S]161g heavy chain V region - Mouse H37? 311
] 533419 heavy chain V region ? Mouse HH?4O
] SJ72 T'cl'll receptor beta cha,n precyrsor V region (ATlI2'2) . Human
3 S435 Ig heavy chain V region - Mouse HleS? 401
3 SS79 Ig heavy chain V region - Mouse H]7-14
35603 Ig IMnbdl?2 chain e region - Ral
3.5666 J9 heavy chain V region - Moust 8 I? , henEallve slquence)
] 5709ll11ary glyCoprotein I- Human
] 5741 Nonspecific cross -reacting antigen precursor - Human
35115 Ig epsilon chain e region? Human
3.5115 Ig epsilon chain e rec,kln ' Human
3.5194 Neur.1 cell adheskln ptoteln precursor? Mouse
3.5912 Ig bppa chain V region - Mouse H]7-60
35971 Ig kilppa chain precursor'll region - Rat IR2
36020 Ig kappa chain V region? Mous. IF6
] 6020 fg kappa chilin V region? Mouse 3010
36027 T'cell receptor beell chain V region (K~ATU - Human
36071 19 heavy chlln V region ? Mouse HP20
36071 Ig heavy chain V regktn ? Mouse HP25
1.6120 T-cl'll receptor alptta ch.ln V regIOn (5c.en - Mouse
36'20 T-cell receptor alptta ch.'n V region (U~ ? Mouse
3.6120 T-cell receptor alph. ch.ln Pfecursor V region (214) ? Mouse
3.6120 T-cell receptor alptt. chain ptecursor V region (4.e]) . Mouse
] 6120 T?cell rec.pc:or Ilptt. chilin precursor'll reqlon (810) ? Mouse
].6302 HLA class 1/ hlSloCompatlb. antigen OX alpha chain prKursor - Human
] 6302 HLA class .. hlstocompatlb. anlAgen. OQ alph. chain precursor' Humiln
36461 T-ul! receptor alptta chilln precursor V region (HAPSIij - Human
] 646S Ig kappa chain precursor V ch.'n - Moys. s.e,-b
36539 Heur.1 un adhesion ptotetn precursor? Mouse
3.6636Ig huvy chain V region - Mouse BI -&'VI1V2 (untatlve slquence)
] 6771 Ig kappa chain precursor V-HI regAon - Human SU?OHl?6
36791 Ig kappa chain V region - Mouse H Ia-S415
36&)J Myelln-assoclatld gtvc:optoteln 11236 tong form ptecursor ? Rae:
3.6&)) Myelln'ilSsoc"tld g~op,oteln IB236 shon form prKursor . Rat
3.6&)3 Myelln-MsOClatld g~oproteln precursor. brain ~ Rar
].6&)J Myebn?assocl.r:. lar,. gtyc:oproteln precursor ? Rat
3.7102 It kappa chain V-III 'eglon - HUman C8
3.7170 Ig kappa chain V-I regIon ' HUman WII(2)
] 7341 Ig lambdl chain e region' Chicken
] 7505 Ig hppa chain precursor V?I region? Human Natm-6
] 75351g heavy chain precursor V regIOn - Mouse 129
] 7600 Ig lambda?5 chain C region - Mouse
3.7779 19 h~avy chain V reg60n - Mouse HP 12
] 790719 kappa chain V region 30S precursor - Humiln
] 790719 kappa chain precursor '1,111- Human Nalm-6
] 7909 19 heavy chain V region? Mouse HP21
] &017 Ntural cell adhHk)n proUtn precursor' Mouse
] 81 ao Ig mu chain e rtglon. b allele? Mouse
3824719 epSilon chain C region - Human
3 8247 ~ epsilon chilln e region - Human
3 &440 ~ kilppa chll" precursor V region? Mouse MAkH
3867119 klppa chain precursor II region? Rat IRI62
1In_
427
428
Bengio, Bengio, Pouliot and Agin
Table 2: Efficiency of detection for some Ig superfamily proteins present in NEW. Mean scores of recognized Ig domains
for each protein type are listed. Recognition efficiency is calculated by dividing the number of proteins correctly identified
(Le., bearing at least one Ig domain) by the total number of
proteins identified by their file description as containing an Ig
domain, multiplied by 100. Numbers in parentheses indicate
the number of complete protein sequences of each type for
each species. All complete sequences for light and heavy immunoglobulin chains of human and mouse origin were
scanned. The threshold was set at 3.0. ND: not done.
Protein
Immunoglobulins,
mouse,
all forms
Immunoglobulins,
human,
all forms
H-2 class II,
all forms
HLA class II,
all forms
T-cell receptor
chains,
mouse,
all forms
T-cell receptor
chains,
human,
all forms
Mean score of
detected domains
(max 4.00)
3.50
Recognition emciency for
Ig-bearing proteins
(see le2end)
98.2 % (55)
3.48
93.8 % (16)
3.33
ND
3.36
ND
3.32
ND
3.41
ND
The vast majority of proteins which scored above 3.0 were of human, mouse,
rat or rabbit origin. A few viral and insect proteins also scored above the
threshold. All proteins in the training set and present in either the NEW or
PROTEIN databases were detected. Proteins detected in the NEW database
are listed in Table I and sorted according to score. Even though only human
MHC class I and II were included in the training set, both mouse H-2 class I
and II were detected. Bovine and rat transplantation antigens were also
detected. These proteins are homologs of human MHC's. For proteins which
include more than one Ig domain contiguously arranged (e.g., carcinoembryonic antigen), all domains were detected if they were sufficiently well conserved. However, domains lacking a feature or possessing a degenerate
feature scored much lower (usually below 3.0) such that they are not recognized when using a threshold value of 3. Recognition of human and mouse immunoglobulin sequences was used to measure recognition efficiency. The rate
of false negatives for immunoglobulins was very low for both species (Table
II). Table III lists the 13 proteins categorized as false positives detected when
searching with a threshold of 3.0. Relative to the total number of domains
detected, this corresponds to a false positive rate of 6.8%. In the strict sense
some of these proteins are not false positives because they do exhibit the expected features of the Ig domain in the correct order. However, inter-feature
A Neural Network to Detect Homologies in Proteins
distances for these pseudo-domains are very different from those observed in
bona fide Ig domains. Proteins which are rich in ,B-sheets, such as rat sodium
channel II and fruit-fly NADH-ubiquinone oxidoreductase chain 1 are also
abundant among the set of false positives. This is not surprising since the Ig
domain is composed of ,B-strands. One solution to this problem lies in the use
of a larger training set as well as the addition of a more intelligent second
stage designed to evaluate inter-feature distances so as to increase the specificity of detection.
Table 3: False positives obtained when searching NEW with a threshold of
3.0. Proteins categorized as false positives are listed. See text for details.
3.0244 Kinase-related transforming protein (src) (Ee 2.7.1.-)
3.0409 Granulocyte-macrophage colony-stimulating
3.0492 NADH-ubiquinone oxidoreductase (Ee 1.6.5.3), chain 5
3.0508 NADH-ubiquinone oxidoreductase (Ee 1.6.5.3), chain 1
3.0561 Protein-tyrosine kinase (Ee 2.7.1.112), lymphocyte - Mouse
3.1931 Hypothetical protein HQLF2 - Cytomegalovirus (strain AD169)
3.2041 Sodium channel protein II - Rat
3.2147 SURF-1 protein - Mouse
3.3039 X-linked chronic granulomatous disease protein - Human
3.4840 Peroxidase (Ee 1.11.1.7) precursor - Human
3.4965 Notch protein - Fruit fly
3.4965 Notch protein - Fruit fly
3.5035 Alkaline phosphatase (EC 3.1.3.1) precursor - Human
5 DISCUSSION
The detection of specific protein domains is becoming increasingly important
since many proteins are constituted of a succession of domains. Unfortunately, domains (Ig or otherwise) are often only weakly homologous with each
other. We have designed a neural network to detect proteins which comprise
Ig domains to evaluate this approach in helping to solve this problem. Alternatives to neural network-based search programs exist. Search programs can
be designed to recognize the flanking Cys-termini regions to the exclusion of
other domain features since these flanks are the best conserved features of Ig
domains (c/. Wang et ai., 1989). However, even Cys-termini can exhibit poor
overall homology and therefore generate statistically insignificant homology
scores when analyzed with the ALIGN program (NBRF) (cf. Williams and
Barclay, 1987). Other search programs (such as Profile Analysis) cannot efficiently handle the large variations in domain size exhibited by the Ig domain
(mostly comprised between 45 and 70 residues). Search results become corrupted by high rates of false positives and negatives. Since the size of the
NBRF protein databases increases considerably each year, the problem of
false positives promises to become crippling if these rates are not substantially
decreased. In view of these problems we have found the application of a
neural network to the detection of Ig domains to be an advantageous solution.
As the state of biological knowledge advances, new Ig domains can be added
to the training set and training resumed. They can learn the statistical features
429
430
Bengio, Bengio, Pouliot and Agio
of the conserved subregions that permit detection of an Ig domain and generalize to new examples of this domain that have a similar distribution. Previously unrecognized and possibly degenerate homologous sequences are therefore likely to be detected.
Acknowledgments
This research was supported by a grant from the Canadian Natural Sciences
and Engineering Research Council to Y.B. We thank CISTI for graciously allowing us access to their experimental BIOMOLE' system.
References
Bengio Y., Cardin R., De Mori R., Merlo E. (1989) Programmable execution
of multi-layered networks for automatic speech recognition, Communications
of the Association for Computing Machinery, 32 (2).
Bengio Y., Cardin R., De Mori R., (1990), Speaker independent speech
recognition with neural networks and speech knowledge, in D.S. Touretzky
(ed.), Advances in Neural Networks Information Processing System,s 2
Blaschuk O.W., Pouliot Y., Holland P.C., (1990). Identification of a conserved region common to cadherins and influenza strain A hemagglutinins. J.
Molec. Biology, 1990, in press.
Devereux, J., Haeberli, P. and Smithies, O. (1984) A comprehensive set of
sequence analysis programs for the VAX. Nucl. Acids Res. 12, 387-395.
Gribskov, M., McLachlan, M., and Eisenber, D. (1987) Profile analysis:
Detection of distantly related proteins. Proc. Natl. Acad. Sci. USA,
84 :4355-4358.
Needleman, S. B. and Wunsch, C. D. (1970) A general method applicable to
the search for similarities in the amino acid sequence of two proteins. J. Mol.
Bioi. 48, 443-453.
Qian, N. and Sejnowski, T. J. (1988) Predicting the secondary structure of
globular proteins using neural network models. J. Mol. Bioi. 202, 865-884.
Rumelhart D.E., Hinton G.E. & Williams R.J. (1986) Learning internal
representation by error propagation. Parallel Distributed Processing, Vol. 1,
MIT Press, Cambridge, pp. 318-362.
Smith, T. F. and Waterman, W. S. (1981). Identification of common molecular subsequences. J. Mol. Bioi. 147 , 195-197.
Stormo, G. D., Schneider, T. D., Gold, L. and Ehrenfeucht, A. Use of the
"perceptron" algorithm to distinguish translational initiation sites in E. coli.
Nucl. Acids Res. 10 , 2997-3010.
Wang, H., Wu, J. and Tang, P. (1989) Superfamily expands. Nature, 337, 514.
Wilbur, W. J. and Lipman, D. J. (1983). Rapid similarity searches of nucleic
acids and protein data banks. Proc. Natl. Acad. Sci. USA 80, 726-730.
Williams, A. F. and Barclay, N. A. (1988) The immunoglobulin superfamilydomains for cell surface recognition. Ann. Rev. Immunol., 6, 381-405.
| 214 |@word trial:1 cu:1 version:1 advantageous:1 nd:5 bf:1 cha:5 hu:2 r:3 cla:1 substitution:1 score:10 heur:1 terminus:2 current:1 virus:2 nt:2 surprising:1 si:1 srd:1 designed:6 recept:1 discrimination:1 capitalizes:1 smith:2 tertiary:1 detecting:2 location:1 ohl:1 c6:1 lor:1 h4:1 beta:11 become:2 consists:1 inter:3 thy:1 expected:1 ra:1 rapid:1 p1:7 roughly:1 ry:1 brain:1 multi:1 rem:2 little:2 cpu:2 window:2 precursor:58 vertebrate:1 wilbur:2 aln:1 substantially:1 elm:1 pseudo:1 ooi:1 hypothetical:2 expands:1 growth:1 prot:2 unit:2 ly:1 grant:1 positive:11 engineering:1 local:1 io:1 acad:2 receptor:21 granulocyte:1 reacting:1 ree:1 becoming:1 antigen:7 co:3 programmed:1 bi:1 statistically:1 acknowledgment:1 procedure:1 bell:1 significantly:1 mhc:2 pre:1 eer:1 specificity:1 protein:56 suggest:1 beca:1 ir2:1 cannot:1 layered:1 sheet:2 influence:1 chronic:1 williams:6 nonspecific:1 starting:4 rabbit:1 formulate:1 qian:2 rule:2 wunsch:2 searching:4 handle:2 variation:4 mcgill:2 user:1 programming:1 us:1 samy:1 origin:2 ppa:3 cysteine:1 recognition:10 rumelhart:2 kappa:18 rec:3 q7:1 bec:1 database:10 observed:2 ft:1 fly:6 wang:3 cy:1 region:91 sun:1 eu:2 src:1 disease:1 nellral:1 mu:1 transforming:1 dynamic:1 tyrosine:1 trained:2 weakly:1 upon:7 localization:1 efficiency:3 easily:1 joint:1 po:1 mh:1 various:2 informatique:1 fast:1 sejnowski:2 detected:11 lift:1 avi:1 outside:1 cardin:2 larger:1 solve:1 otherwise:2 transplantation:1 sequence:18 p4:6 aligned:1 gen:1 poorly:1 degenerate:2 gold:1 description:1 afr:1 montreal:4 colony:1 lyc:1 op:2 school:1 strong:1 soc:1 dividing:1 p2:6 subregion:1 indicate:1 uu:1 pii:2 exhibiting:1 correct:2 allele:1 human:66 in_:1 globular:1 rar:1 sor:1 ao:1 drosophila:2 tln:1 biological:2 barclay:5 molec:1 helping:1 sufficiently:2 stormo:2 mo:2 substituting:1 cen:1 consecutive:1 a2:2 recognizer:1 proc:4 applicable:1 bond:1 currently:1 prf:1 council:1 ctl:1 mclachlan:1 mit:1 clearly:1 cr:1 bet:7 og:1 hun:1 derived:1 lon:4 ral:1 detect:8 sense:1 graciously:1 mou:1 i0:1 vl:1 hidden:4 spurious:1 ancestor:1 comprising:2 bovine:1 overall:6 among:1 ill:2 translational:1 insect:2 ell:7 comprise:1 lipman:2 biology:2 identical:1 s22:1 alter:1 distantly:1 yoshua:1 intelligent:1 few:2 composed:1 gamma:2 national:1 recognize:2 comprehensive:1 ehrenberg:1 detection:6 fd:1 a5:1 highly:1 rae:1 violation:1 analyzed:1 pc:1 light:1 tj:1 beu:2 natl:2 chain:105 iiy:1 kt:1 necessary:1 machinery:1 re:3 closs:2 abundant:1 immunoglobulin:9 ollt:2 resumed:1 pia:1 lymphocyte:2 comprised:1 a2m:2 gr:1 characterize:1 scanning:2 corrupted:1 ido:1 considerably:1 clone:1 discriminating:1 ie:1 bu:1 nma:1 mouse:81 von:1 containing:1 hn:1 possibly:1 dr:2 lambda:1 bela:3 stochastically:1 coli:1 tft:2 til:2 de:4 explicitly:2 ad:1 stream:1 view:1 linked:1 sup:2 parallel:1 minimize:1 il:1 acid:13 characteristic:2 efficiently:2 succession:1 correspond:1 identify:1 crippling:1 ant:6 generalize:1 weak:1 identification:2 published:1 suffers:1 touretzky:1 ed:4 evaluates:1 pp:1 involved:1 universite:1 rld:1 knowledge:2 spt:2 back:2 ta:3 arranged:1 evaluated:1 though:2 ox:2 done:1 biomedical:1 stage:6 ei:1 su:1 a7:1 o:2 propagation:3 lar:1 indicated:1 name:1 effect:1 homologs:1 usa:2 homology:16 needleman:2 hence:1 chemical:2 ehrenfeucht:1 ind:1 ll:10 noted:1 speaker:1 rat:14 eii:6 complete:3 tn:1 possessing:1 common:3 viral:1 mt:1 ji:1 macrophage:1 influenza:1 association:1 significant:1 s5:1 cambridge:1 ai:1 automatic:1 hp:1 immune:1 mlb:2 pte:1 access:1 entail:1 similarity:3 surface:5 etc:1 patrick:1 align:1 exclusion:1 moderate:1 certain:1 initiation:1 success:1 fen:1 conserved:13 cham:5 filename:1 greater:1 additional:1 schneider:1 recognized:3 ii:26 h3a:1 characterized:2 cross:2 long:1 concerning:1 equally:1 molecular:1 parenthesis:1 impact:1 das:1 j3:1 cell:21 ion:2 chicken:1 addition:1 residue:8 adhesion:2 decreased:1 eliminates:1 posse:2 exhibited:2 probably:1 file:2 duplication:1 strict:1 member:2 oq:2 ee:8 smote:1 presence:1 feedforward:1 bengio:13 iii:2 canadian:1 affect:1 architecture:1 identified:4 intensive:1 whether:1 handled:1 notch:2 ul:2 penalty:2 speech:4 hia:1 constitute:1 repeatedly:1 bli:1 programmable:1 generally:1 myelin:1 se:1 listed:6 j9:1 amount:1 hla:15 subregions:5 generate:1 exist:1 vy:1 correctly:1 biq:1 promise:1 vol:1 group:6 four:4 threshold:9 cutoff:3 vast:2 contiguously:1 sum:1 year:1 run:2 c02:1 wu:1 p3:5 fide:2 comparable:1 layer:3 bound:1 hi:3 fl:1 distinguish:1 scanned:3 constraint:3 chair:1 c8:1 relatively:2 department:1 designated:1 according:2 tv:1 neur:1 poor:1 membrane:1 increasingly:1 character:1 departement:2 modification:1 rev:1 hl:2 pr:6 flanking:1 computationally:1 ln:15 mori:2 previously:2 hh:2 ge:1 lyt:1 hap:1 wii:2 permit:1 multiplied:1 prec:2 glyc:1 occurrence:1 cia:3 gribskov:2 alternative:1 existence:1 kln:1 cf:2 include:1 cryptic:1 epsilon:8 added:1 hum:5 rt:2 traditional:1 nr:1 exhibit:3 dp:1 distance:6 thank:1 sci:2 eif:1 majority:2 length:3 furthemore:1 lg:2 unfortunately:1 mostly:1 fe:1 negative:5 unrecognized:1 bona:2 design:3 kinase:3 peroxidase:1 allowing:1 nucleic:1 waterman:2 pat:1 hinton:1 communication:1 strain:2 lb:1 canada:1 introduced:1 bk:2 cys:2 required:2 c3:1 optimized:1 huma:1 hour:2 alph:4 assembled:1 address:1 below:3 pattern:1 usually:1 chal:1 program:16 built:1 including:1 rf:1 max:1 ia:2 critical:2 natural:1 rely:1 homologous:3 predicting:1 force:1 nucl:2 sodium:4 representing:1 hll:1 mn:1 vax:2 igen:1 coupled:1 text:2 review:1 merlo:1 relative:1 wisconsin:1 lacking:1 proportional:1 localized:2 foundation:1 degree:1 fruit:4 tld:1 bank:1 heavy:26 genetics:1 elsewhere:1 accounted:1 supported:1 lln:1 allow:1 perceptron:2 institute:1 superfamily:5 fg:1 distributed:1 celf:1 calculated:1 valid:1 ending:4 rich:1 ig:66 ec:4 eglo:1 melanogaster:1 alpha:7 implicitly:1 e1i:1 assumed:1 uls:1 subsequence:3 search:14 un:1 table:6 aloha:1 nature:1 learn:1 channel:4 glycoprotein:4 ku:1 mncn:1 mol:3 flank:1 bearing:2 necessarily:1 cl:2 domain:55 eukaryotic:1 surf:2 constituted:1 scored:3 profile:5 ceu:1 amino:10 categorized:2 site:1 en:4 ff:1 tong:1 position:4 col:3 lie:1 learns:1 tang:1 specific:2 list:1 insignificant:1 incorporating:1 false:14 execution:1 gap:1 suited:1 fc:1 remi:1 likely:1 clas:4 tch:1 strand:2 yannick:1 neurological:1 shon:2 holland:2 collectively:1 ch:20 corresponds:2 iph:2 stimulating:1 bioi:3 sorted:2 kli:1 ann:1 man:1 considerable:1 change:1 included:1 except:1 total:2 specie:2 secondary:1 accepted:1 experimental:1 perceptrons:1 hies:1 internal:1 searched:1 scan:2 evaluate:3 reg:1 |
1,252 | 2,140 | Learning Attractor Landscapes for
Learning Motor Primitives
Auke Jan Ijspeert1,3?, Jun Nakanishi2 , and Stefan Schaal1,2
University of Southern California, Los Angeles, CA 90089-2520, USA
2
ATR Human Information Science Laboratories, Kyoto 619-0288, Japan
3
EPFL, Swiss Federal Institute of Technology, Lausanne, Switzerland
[email protected], [email protected], [email protected]
1
Abstract
Many control problems take place in continuous state-action spaces,
e.g., as in manipulator robotics, where the control objective is often defined as finding a desired trajectory that reaches a particular
goal state. While reinforcement learning offers a theoretical framework to learn such control policies from scratch, its applicability to
higher dimensional continuous state-action spaces remains rather
limited to date. Instead of learning from scratch, in this paper we
suggest to learn a desired complex control policy by transforming
an existing simple canonical control policy. For this purpose, we
represent canonical policies in terms of differential equations with
well-defined attractor properties. By nonlinearly transforming the
canonical attractor dynamics using techniques from nonparametric
regression, almost arbitrary new nonlinear policies can be generated without losing the stability properties of the canonical system. We demonstrate our techniques in the context of learning a
set of movement skills for a humanoid robot from demonstrations
of a human teacher. Policies are acquired rapidly, and, due to the
properties of well formulated differential equations, can be re-used
and modified on-line under dynamic changes of the environment.
The linear parameterization of nonparametric regression moreover
lends itself to recognize and classify previously learned movement
skills. Evaluations in simulations and on an actual 30 degree-offreedom humanoid robot exemplify the feasibility and robustness
of our approach.
1
Introduction
Learning control is formulated in one of the most general forms as learning a control
policy u = ?(x, t, w) that maps a state x, possibly in a time t dependent way, to an
action u; the vector w denotes the adjustable parameters that can be used to optimize the policy. Since learning control policies (CPs) based on atomic state-action
representations is rather time consuming and faces problems in higher dimensional
and/or continuous state-action spaces, a current topic in learning control is to use
?
http://lslwww.epfl.ch/?ijspeert/
higher level representations to achieve faster and more robust learning [1, 2]. In this
paper we suggest a novel encoding for such higher level representations based on the
analogy between CPs and differential equations: both formulations suggest a change
of state given the current state of the system, and both usually encode a desired
goal in form of an attractor state. Thus, instead of shaping the attractor landscape
of a policy tediously from scratch by traditional methods of reinforcement learning,
we suggest to start out with a differential equation that already encodes a rough
form of an attractor landscape and to only adapt this landscape to become more
suitable to the current movement goal. If such a representation can keep the policy
linear in the parameters w, rapid learning can be accomplished, and, moreover, the
parameter vector may serve to classify a particular policy.
In the following sections, we will first develop our learning approach of shaping attractor landscapes by means of statistical learning building on preliminary previous
work [3, 4].1 Second, we will present a particular form of canonical CPs suitable
for manipulator robotics, and finally, we will demonstrate how our methods can be
used to classify movement and equip an actual humanoid robot with a variety of
movement skills through imitation learning.
2
Learning Attractor Landscapes
We consider a learning scenario where the goal of control is to attain a particular
attractor state, either formulated as a point attractor (for discrete movements) or
as a limit cycle (for rhythmic movements). For point attractors, we require that the
CP will reach the goal state with a particular trajectory shape, irrespective of the
initial conditions ? a tennis swing toward a ball would be a typical example of such
a movement. For limit cycles, the goal is given as the trajectory shape of the limit
cycle and needs to be realized from any start state, as for example, in a complex
drumming beat hitting multiple drums during one period. We will assume that,
as the seed of learning, we obtain one or multiple example trajectories, defined by
positions and velocities over time. Using these samples, an asymptotically stable
CP is to be generated, prescribing a desired velocity given a particular state 2 .
Various methods have been suggested to solve such control problems in the literature. As the simplest approach, one could just use one of the demonstrated
trajectories and track it as a desired trajectory. While this would mimic this one
particular trajectory, and scaling laws could account for different start positions
[5], the resultant control policy would require time as an explicit variable and thus
become highly sensitive toward unforeseen perturbations in the environment that
would disrupt the normal time flow. Spline-based approaches [6] have a similar
problem. Recurrent neural networks were suggested as a possible alternative that
can avoid explicit time indexing ? the complexity of training these networks to obtain stable attractor landscapes, however, has prevented a widespread application
so far. Finally, it is also possible to prime a reinforcement learning system with
sample trajectories and pursue one of the established continuous state-action learning algorithms; investigations of such an approach, however, demonstrated rather
limited efficiency [7]. In the next sections, we present an alternative and surprisingly
simple solution to learning the control problem above.
1
Portions of the work presented in this paper have been published in [3, 4]. We here
extend these preliminary studies with an improvement and simplification of the rhythmic
system, an integrated view of the interpretation of both the discrete and rhythmic CPs, the
fitting of a complete alphabet of Grafitti characters, and an implementation of automatic
allocation of centers of kernel functions for locally weighted learning.
2
Note that we restrict our approach to purely kinematic CPs, assuming that the movement system is equipped with an appropriate feedback and feedforward controller that can
accurately track the kinematic plans generated by our policies.
Table 1: Discrete and Rhythmic control policies. ?z , ?z , ?v , ?v , ?z , ?z , ?, ?i and ci are
positive constants. x0 is the start state of the discrete system in order to allow nonzero initial conditions. The design parameters of the discrete system are ? , the temporal
scaling factor, and g, the goal position. The design parameters of the rhythmic system
are ym , the baseline of the oscillation, ? , the period divided by 2?, and ro , the amplitude
of oscillations. The parameters wi are fitted to a demonstrated trajectory using Locally
Weighted Learning.
Discrete
PN
?
?i wiT v
i=1
? y? = z + P
N
i=1
?i
? z? = ?z (?z (g ? y) ? z)
? = [v]
v
? v? = ?v (?v (g ? x) ? v)
? x? = v ?
?
2
0
?i = exp ?hi ( x?x
g?x0 ? ci )
ci ? [0, 1]
2.1
Rhythmic
PN
?
?i wiT v
i=1
? y? = z + P
N
i=1
?i
? z? = ?z (?z (ym ? y) ? z)
? = [r cos ?, r sin ?]T
v
? ?? = 1
? r? = ??(r ? r0 )
?
?
?i = exp ?hi (mod(?, 2?) ? ci )2
ci ? [0, 2?]
Dynamical systems for Discrete Movements
Assume we have a basic control policy (CP), for instance, instantiated by the second
order attractor dynamics
? z? = ?z (?z (g ? y) ? z)
? y? = z + f
(1)
where g is a known goal state, ?z , ?z are time constants, ? is a temporal scaling
factor (see below) and y, y? correspond to the desired position and velocity generated
by the policy as a movement plan. For appropriate parameter settings and f = 0,
these equations form a globally stable linear dynamical system with g as a unique
point attractor. Could we insert a nonlinear function f in Eq.1 to change the rather
trivial exponential convergence of y to allow more complex trajectories on the way
to the goal? As such a change of Eq.1 enters the domain of nonlinear dynamics,
an arbitrary complexity of the resulting equations can be expected. To the best
of our knowledge, this has prevented research from employing generic learning in
nonlinear dynamical systems so far. However, the introduction of an additional
canonical dynamical system (x, v)
? v? = ?v (?v (g ? x) ? v)
and the nonlinear function f
PN
i=1 ?i wi v
f (x, v, g) = P
N
i=1 ?i
? x? = v
(2)
?
?
?i = exp ?hi (x/g ? ci )2
(3)
can alleviate this problem. Eq.2 is a second order dynamical system similar to
Eq.1, however, it is linear and not modulated by a nonlinear function, and, thus,
its monotonic global convergence to g can be guaranteed with a proper choice of
?v and ?v . Assuming that all initial conditions of the state variables x, v, y, z are
initially zero, the quotient x/g ? [0, 1] can serve as a phase variable to anchor the
Gaussian basis functions ?i (characterized by a center ci and bandwidth hi ), and v
can act as a ?gating term? in the nonlinear function (3) such that the influence of
this function vanishes at the end of the movement. Assuming boundedness of the
weights wi in Eq.3, it can be shown that the combined dynamical system (Eqs.1?3)
asymptotically converges to the unique point attractor g.
Given that f is a normalized basis function representation with linear parameterization, it is obvious that this choice of a nonlinearity allows applying a variety of
?1
?5
0
0.5
1
1.5
2
?10
0
0.5
1
1.5
?3
2
0
mod(?,2?)
0
0.5
1
1.5
0.5
1
1.5
Time [s]
2
0.5
1
1.5
?40
2
0
0.5
1
1.5
0
2
1.5
0.5
0.5
0
0.5
1
1.5
2
1
1.5
2
0
0.5
1
1.5
2
0
0.5
1
1.5
2
0.5
1
0
0.5
2
1
1
0
1
4
0
2
0.5
0
0
2
r cos(?)
1
0
?20
6
?i
x
v
2
20
?2
0.5
0
40
0
?1
1
3
1
r sin(?)
0
y
0
dy/dt
5
1
?i
10
2
dy/dt
y
3
0
?0.5
?1
0
Time [s]
0.5
1
Time [s]
1.5
2
0
?0.5
?1
Time [s]
Figure 1: Examples of time evolution of the discrete CPs (left) and rhythmic CPs (right).
The parameters wi have been adjusted to fit y? demo (t) = 10 sin(2?t) exp(?t2 ) for the discrete CPs and y? demo (t) = 2? cos(2?t) ? 6? sin(6?t) for the rhythmic CPs.
learning algorithms to find the wi . For learning from a given sample trajectory,
characterized by a trajectory ydemo (t), y? demo (t) and duration T , a supervised learning problem can be formulated with the target trajectory ftarget = ? y? demo ? zdemo
for Eq.1 (right), where zdemo is obtained by integrating Eq.1 (left) with ydemo instead of y. The corresponding goal state is g = ydemo (T ) ? ydemo (t = 0), i.e., the
sample trajectory was translated to start at y = 0. In order to make the nominal
(i.e., assuming f = 0) dynamics of Eqs.1 and 2 span the duration T of the sample
trajectory, the temporal scaling factor ? is adjusted such that the nominal dynamics
achieves 95% convergence at t = T . For solving the function approximation problem, we chose a nonparametric regression technique from locally weighted learning
(LWL) [8] as it allows us to determine the necessary number of basis functions, their
centers ci , and bandwidth hi automatically ? in essence, for every basis function
?i , LWL performs a locally weighted regression of the training data to obtain an
approximation of the tangent of the function to be approximated within the scope of
the kernel, and a prediction for a query point is achieved by a ?i -weighted average
of the predictions all local models. Moreover, as will be explained later, the parameters wi learned by LWL are also independent of the number of basis functions,
such that they can be used robustly for categorization of different learned CPs.
In summary, by anchoring a linear learning system with nonlinear basis functions in
the phase space of a canonical dynamical system with guaranteed attractor properties,
we are able to learn complex attractor landscapes of nonlinear differential equations
without losing the asymptotic convergence to the goal state.
2.2
Extension to Limit Cycle Dynamics
The system above can be extended to limit cycle dynamics by replacing the canonical system (x, v) with, for instance, the following rhythmic system which has a
stable limit cycle in terms of polar coordinates (?, r):
? ?? = 1
? r? = ??(r ? r0 )
(4)
Similar to the discrete system, the rhythmic canonical system serves to provide
? = [r cos ?, r sin ?]T and phase variable mod(?, 2?) to
both an amplitude signal v
the basis function ?i of the control policy (z, y):
PN
T?
i=1 ?i wi v
(5)
? z? = ?z (?z (ym ? y) ? z)
? y? = z + P
N
i=1 ?i
where ym is an anchor point for the oscillatory trajectory. Table 1 summarizes the
proposed discrete and rhythmic CPs, and Figure 1 shows exemplary time evolutions
of the complete systems.
2.3
Special Properties of Control Policies based on Dynamical Systems
Spatial and Temporal Invariance An interesting property of both discrete and
rhythmic CPs is that they are spatially and temporally invariant. Scaling of the goal
g for the discrete CP and of the amplitude r0 for the rhythmic CP does not affect the
topology of the attractor landscape. Similarly, the period (for the rhythmic system)
and duration (for the discrete system) of the trajectory y is directly determined
by the parameter ? . This means that the amplitude and durations/periods of
learned patterns can be independently modified without affecting the qualitative
shape of trajectory y. In section 3, we will exploit these properties to reuse a
learned movement (such as a tennis swing, for instance) in novel conditions (e.g
toward new ball positions).
Robustness against Perturbations When considering applications of our approach to physical systems, e.g., robots and humanoids, interactions with the environment may require an on-line modification of the policy. An obstacle can, for
instance, block the trajectory of the robot, in which case large discrepancies between
desired positions generated by the control policy and actual positions of the robot
will occur. As outlined in [3], the dynamical system formulation allows feeding back
an error term between actual and desired positions into the CPs, such that the time
evolution of the policy is smoothly paused during a perturbation, i.e., the desired
position y is modified to remain close to the actual position y?. As soon as the
perturbation stops, the CP rapidly resumes performing the (time-delayed) planned
trajectory. Note that other (task-specific) ways to cope with perturbations can be
designed. Such on-line adaptations are one of the most interesting properties of
using autonomous differential equations for CPs.
Movement Recognition Given the temporal and spatial invariance of our policy
representation, trajectories that are topologically similar tend to be fit by similar parameters wi , i.e., similar trajectories at different speeds and/or different amplitudes
will result in similar wi . In section 3.3, we will use this property to demonstrate
the potential of using the CPs for movement recognition.
3
Experimental Evaluations
3.1 Learning of Rhythmic Control Policies by Imitation
We tested the proposed CPs in a learning by demonstration task with a humanoid
robot. The robot is a 1.9-meter tall 30 DOFs hydraulic anthropomorphic robot
with legs, arms, a jointed torso, and a head [9]. We recorded trajectories performed
by a human subject using a joint-angle recording system, the Sarcos Sensuit (see
Figure 2, top). The joint-angle trajectories are fitted by the CPs, with one CP
per degree of freedom (DOF). The CPs are then used to replay the movement
in the humanoid robot, using an inverse dynamics controller to track the desired
trajectories generated by the CPs. The actual positions y? of each DOF are fed back
into the CPs in order to take perturbations into account.
Using the joint-angle recording system, we recorded a set of rhythmic movements
such as tracing a figure 8 in the air, or a drumming sequence on a bongo (i.e.
without drumming sticks). Six DOFs for both arms were recorded (three at the
shoulder, one at the elbow, and two at the wrist). An exemplary movement and its
replication by the robot is demonstrated in Figure 2 (top). Figure 2 (left) shows the
joint trajectories over one period of an exemplary drumming beat. Demonstrated
and learned trajectories are superposed. For the learning, the base frequency was
extracted by hand such as to provide the parameter ? to the rhythmic CP.
Once a rhythmic movement has been learned by the CP, it can be modulated
in several ways. Manipulating r0 and ? for all DOFs amounts to simultaneously
0.1
0
?0.1
?0.2
R_SFE
1
1.5
2
R_SAA
0
0
0.5
1
1.5
0.5
1
1.5
0.5
1
1.5
0
0.5
0
0.5
1
2
1
1.5
Time [s]
2
1
1.5
2
0
0.5
1
1.5
2
0
0
1
2
3
4
5
6
7
8
9
10
0
1
2
3
4
5
6
7
8
9
10
0
1
2
3
4
5
6
7
8
9
10
0
1
2
3
4
5
Time [s]
6
7
8
9
10
3
2
B
1
0
0
R_WR
0.2
0
?0.2
0.5
0.5
1
1.5
2
0
?0.2
?0.4
?0.6
?0.8
2
2
1
0
0.4
0.2
0
2
1.5
2
A
0.1
0.05
0
?0.05
R_WFE
0
L_WR
0.5
0.15
0.1
0.05
0
R_HR
0.6
0.4
0.2
0
?0.2
?0.4
0
R_EB
L_SFE
L_HR
0.2
0
?0.2
?0.4
L_EB
0.05
0
?0.05
?0.1
?0.15
L_WFE
L_SAA
3
0.05
0
?0.05
?0.1
3
2
C
0
0.5
1
1.5
2
0
0.5
1
1.5
2
0
0.5
1
1.5
Time [s]
2
0.3
0.2
0.1
0
1
0
3
0
?0.2
?0.4
2
D
1
0
Figure 2: Top: Humanoid robot learning a figure-8 movement from a human demonstration. Left: Recorded drumming movement performed with both arms (6 DOFs per
arm). The dotted lines and continuous lines correspond to one period of the demonstrated and learned trajectories, respectively. Right: Modification of the learned rhythmic pattern (flexion/extension of the right elbow, R EB). A: trajectory learned by the
rhythmic CP, B: temporary modification with r?0 = 2r0 , C: ?? = ? /2, D: y?m = ym + 1
(dotted line), where r?0 , ??, and y?m correspond to modified parameters between t=3s
and t=7s. Movies of the human subject and the humanoid robot can be found at
http://lslwww.epfl.ch/?ijspeert/humanoid.html.
modulate the amplitude and period of all DOFs, while keeping the same phase
relation between DOFs. This might be particularly useful for a drumming task
in order to replay the same beat pattern at different speeds and/or amplitudes.
Alternatively, the r0 and ? parameters can be modulated independently for the
DOFs each arm, in order to be able to change the beat pattern (doubling the
frequency of one arm, for instance). Figure 2 (right) illustrates different modulations
which can be generated by the rhythmic CPs. For reasons of clarity, only one DOF
is showed. The rhythmic CP can smoothly modulate the amplitude, frequency, and
baseline of the oscillations.
3.2 Learning of Discrete Control Policies by Imitation
In this experiment, the task for the robot was to learn tennis forehand and backhand
swings demonstrated by a human wearing the joint-angle recording system. Once
a particular swing has been learned, the robot is able to repeat the swing motion
to different cartesian targets, by providing new goal positions g to the CPs for the
different DOFs. Using a system of two-cameras, the position of the ball is given
to an inverse kinematic algorithm which computes these new goals in joint space.
When the new ball positions are not too distant from the original cartesian target,
the modified trajectories reach the ball with swing motions very similar to those
used for the demonstration.
3.3 Movement Recognition using the Discrete Control Policies
Our learning algorithm, Locally Weighted Learning [8], automatically sets the number of the kernel functions and their centers ci and widths hi depending on the complexity of the function to be approximated, with more kernel functions for highly
Figure 3: Humanoid robot learning a forehand swing from a human demonstration.
nonlinear details of the movement. An interesting aspect of locally weighted regression is that the regression parameters wi of each kernel i do not depend on the
other kernels, since regression is based on a separate cost function for each kernel.
This means that kernel functions can be added or removed without affecting the
parameters wi of the other kernels.
We here use this feature to perform movement recognition within a large variety
of trajectories, based on a small subset of kernels at fixed locations c i in phase
space. These fixed kernels are common for fitting all the trajectories, in addition
to the kernels automatically added by the LWL algorithm. The stability of their
parameters wi w.r.t. other kernels generated by LWL makes them well-suited for
comparing qualitative trajectory shapes.
To illustrate the possibility of using the CPs for movement recognition (i.e., recognition of spatiotemporal patterns, not just spatial patterns as in traditional character
recognition), we carried out a simple task of fitting trajectories performed by a human user when drawing two-dimensional single-stroke patterns. The 26 letters of
the Graffiti alphabet used in hand-held computers were chosen. These characters
are drawn in a single stroke, and are fed as a two-dimensional trajectory (x(t), y(t))
to be fitted by our system. Five examples of each character were presented (see
Figure 4 for four examples).
Fixed sets of five kernels per DOF were set aside for movement recognition. The
wT w
correlation |waa||wbb | between their parameter vectors wa and wb of character a and
b can be used to classify movements with similar velocity profiles (Figure 4, right).
For instance, for the 5 instances of the N, I, P, S, characters, the correlation is
systematically higher with the four other examples of the same character. These
similarities in weight space can therefore serve as basis for recognizing demonstrated
movements by fitting them and comparing the fitted parameters wi with those
of previously learned policies in memory. In this example, a simple one-nearestneighbor classifier in weight space would serve the purpose. Using such a classifier
within the whole alphabet (5 instances of each letter), we obtained a 84% recognition
rate (i.e. 110 out of the 130 instances had a highest correlation with an instance of
the same letter). Further studies are required to evaluate the quality of recognition
in larger training and test sets ? what we wanted to demonstrate is the ability
for recognition without any specific system tuning or sophisticated classification
algorithm.
4
Conclusion
Based on the analogy between autonomous differential equations and control policies, we presented a novel approach to learn control policies of basic movement skills
by shaping the attractor landscape of nonlinear differential equations with statistical learning techniques. To the best of our knowledge, the presented approach is the
first realization of a generic learning system for nonlinear dynamical systems that
460
480
440
460
420
440
400
420
480
350
380
380
400
360
380
340
360
16
340
320
200
300
150
320
280
50
100
150
200
250
300
350
400
450
200
250
300
X
350
250
300
350
200
500
400
450
450
450
400
300
0.2
0.4
0.6
0.8
1
1.2
0
0.2
0.4
0.6
400
10
I
8
400
350
350
300
300
0.8
350
Y
Y
Y
Y
400
350
0
300
X
500
250
250
X
X
200
12
400
500
300
14
P
340
200
400
500
100
18
420
Y
360
250
S
440
400
Y
Y
300
20
460
Y
400
0
0.2
0.4
0.6
0.8
1
1.2
6
0
0.2
0.4
0.6
0.8
1
4
400
400
300
250
280
200
0
0.2
0.4
0.6
0.8
1
1.2
Time [s]
260
2
X
X
X
X
300
300
250
N
350
320
300
150
350
340
350
0
0.2
0.4
Time [s]
0.6
0.8
250
0
0.2
0.4
0.6
Time [s]
0.8
1
1.2
200
0
0
0.2
0.4
0.6
Time [s]
0.8
0
2
4
6
8
10
12
14
16
18
20
1
N
I
P
S
Figure 4: Left: Examples of two-dimensional trajectories fitted by the CPs. The demonstrated and fitted trajectories are shown with dotted and continuous lines, respectively.
Right: Correlation between the weight vectors of the 20 characters (5 of each letter) fitted
by the system. The gray scale is proportional to the correlation, with black corresponding
to a correlation of +1 (max. correlation) and white to a correlation of 0 or smaller.
can guarantee basic stability and convergence properties of the learned nonlinear
systems. We demonstrated the applicability of the suggested techniques by learning various movement skills for a complex humanoid robot by imitation learning,
and illustrated the usefulness of the learned parameterization for recognition and
classification of movement skills. Future work will consider (1) learning of multidimensional control policies without assuming independence between the individual
dimensions, and (2) the suitability of the linear parameterization of the control
policies for reinforcement learning.
Acknowledgments
This work was made possible by support from the US National Science Foundation (Awards
9710312 and 0082995), the ERATO Kawato Dynamic Brain Project funded by the Japan
Science and Technology Corporation, the ATR Human Information Science Laboratories,
and Communications Research Laboratory (CRL).
References
[1] R. Sutton and A.G. Barto. Reinforcement learning: an introduction. MIT Press, 1998.
[2] F.A. Mussa-Ivaldi. Nonlinear force fields: a distributed system of control primitives for
representing and learning movements. In IEEE International Symposium on Computational Intelligence in Robotics and Automation, pages 84?90. IEEE, Computer Society,
Los Alamitos, 1997.
[3] A.J. Ijspeert, J. Nakanishi, and S. Schaal. Movement imitation with nonlinear dynamical systems in humanoid robots. In IEEE International Conference on Robotics and
Automation (ICRA2002), pages 1398?1403. 2002.
[4] A.J. Ijspeert, J. Nakanishi, and S. Schaal. Learning rhythmic movements by demonstration using nonlinear oscillators. In Proceedings of the IEEE/RSJ Int. Conference
on Intelligent Robots and Systems (IROS2002), pages 958?963. 2002.
[5] S. Kawamura and N. Fukao. Interpolation for input torque patterns obtained through
learning control. In Proceedings of The Third International Conference on Automation,
Robotics and Computer Vision (ICARCV?94). 1994.
[6] H. Miyamoto, S. Schaal, F. Gandolfo, Y. Koike, R. Osu, E. Nakano, Y. Wada, and
M. Kawato. A kendama learning robot based on bi-directional theory. Neural Networks,
9:1281?1302, 1996.
[7] S. Schaal. Learning from demonstration. In M. C. Mozer, M. Jordan, and T. Petsche,
editors, Advances in Neural Information Processing Systems 9, pages 1040?1046. Cambridge, MA, MIT Press, 1997.
[8] S. Schaal and C.G. Atkeson. Constructive incremental learning from only local information. Neural Computation, 10(8):2047?2084, 1998.
[9] C. G. Atkeson, J. Hale, M. Kawato, S. Kotosaka, F. Pollick, M. Riley, S. Schaal,
S. Shibata, G. Tevatia, A. Ude, and S. Vijayakumar. Using humanoid robots to study
human behavior. IEEE Intelligent Systems, 15:46?56, 2000.
| 2140 |@word simulation:1 boundedness:1 ivaldi:1 initial:3 existing:1 current:3 comparing:2 distant:1 shape:4 motor:1 wanted:1 designed:1 aside:1 intelligence:1 parameterization:4 sarcos:1 location:1 five:2 differential:8 become:2 symposium:1 replication:1 qualitative:2 fitting:4 x0:2 acquired:1 expected:1 rapid:1 behavior:1 brain:1 anchoring:1 torque:1 globally:1 automatically:3 actual:6 equipped:1 considering:1 elbow:2 project:1 moreover:3 what:1 pursue:1 finding:1 corporation:1 guarantee:1 temporal:5 every:1 multidimensional:1 act:1 ro:1 classifier:2 stick:1 control:27 positive:1 local:2 limit:6 sutton:1 encoding:1 modulation:1 interpolation:1 might:1 chose:1 black:1 eb:1 nearestneighbor:1 lausanne:1 co:5 limited:2 bi:1 acknowledgment:1 unique:2 camera:1 wrist:1 atomic:1 block:1 swiss:1 jan:1 attain:1 integrating:1 suggest:4 close:1 pollick:1 context:1 influence:1 applying:1 superposed:1 optimize:1 map:1 demonstrated:10 center:4 primitive:2 duration:4 independently:2 wit:2 his:1 stability:3 coordinate:1 autonomous:2 target:3 nominal:2 user:1 losing:2 velocity:4 approximated:2 recognition:12 particularly:1 enters:1 cycle:6 movement:34 removed:1 highest:1 mozer:1 transforming:2 environment:3 complexity:3 vanishes:1 dynamic:10 depend:1 solving:1 serve:4 purely:1 efficiency:1 basis:8 translated:1 joint:6 various:2 hydraulic:1 alphabet:3 instantiated:1 query:1 dof:4 larger:1 solve:1 drawing:1 ability:1 itself:1 sequence:1 exemplary:3 interaction:1 adaptation:1 realization:1 date:1 rapidly:2 backhand:1 achieve:1 los:2 convergence:5 categorization:1 incremental:1 converges:1 tall:1 depending:1 develop:1 recurrent:1 illustrate:1 eq:9 quotient:1 switzerland:1 human:10 require:3 kawamura:1 feeding:1 offreedom:1 preliminary:2 investigation:1 alleviate:1 anthropomorphic:1 suitability:1 adjusted:2 insert:1 extension:2 normal:1 exp:4 seed:1 scope:1 sschaal:1 achieves:1 purpose:2 polar:1 sensitive:1 weighted:7 stefan:1 federal:1 rough:1 mit:2 gaussian:1 modified:5 rather:4 avoid:1 pn:4 barto:1 forehand:2 encode:1 schaal:6 improvement:1 baseline:2 dependent:1 epfl:3 prescribing:1 integrated:1 initially:1 relation:1 manipulating:1 shibata:1 classification:2 html:1 plan:2 spatial:3 special:1 field:1 once:2 mimic:1 discrepancy:1 t2:1 spline:1 future:1 intelligent:2 simultaneously:1 recognize:1 national:1 individual:1 delayed:1 usc:2 phase:5 mussa:1 attractor:19 freedom:1 highly:2 kinematic:3 possibility:1 evaluation:2 held:1 necessary:1 dofs:8 desired:10 re:1 miyamoto:1 theoretical:1 fitted:7 instance:10 classify:4 obstacle:1 wb:1 planned:1 riley:1 applicability:2 cost:1 wada:1 subset:1 usefulness:1 recognizing:1 too:1 teacher:1 spatiotemporal:1 combined:1 international:3 vijayakumar:1 ym:5 unforeseen:1 recorded:4 possibly:1 japan:2 account:2 potential:1 automation:3 int:1 later:1 view:1 performed:3 portion:1 start:5 gandolfo:1 air:1 correspond:3 landscape:10 resume:1 directional:1 koike:1 accurately:1 trajectory:37 published:1 stroke:2 oscillatory:1 reach:3 against:1 frequency:3 obvious:1 resultant:1 stop:1 exemplify:1 knowledge:2 torso:1 shaping:3 amplitude:8 sophisticated:1 back:2 higher:5 dt:2 supervised:1 formulation:2 kendama:1 just:2 correlation:8 hand:2 replacing:1 nonlinear:16 widespread:1 quality:1 gray:1 manipulator:2 usa:1 building:1 normalized:1 swing:7 evolution:3 spatially:1 laboratory:3 nonzero:1 illustrated:1 white:1 sin:5 during:2 width:1 erato:1 essence:1 waa:1 complete:2 demonstrate:4 performs:1 cp:11 motion:2 lwl:5 novel:3 common:1 kawato:3 physical:1 jp:1 extend:1 interpretation:1 cambridge:1 automatic:1 tuning:1 outlined:1 similarly:1 nonlinearity:1 had:1 funded:1 robot:21 tennis:3 stable:4 similarity:1 base:1 showed:1 prime:1 scenario:1 accomplished:1 additional:1 r0:6 determine:1 period:7 signal:1 multiple:2 kyoto:1 faster:1 adapt:1 characterized:2 offer:1 divided:1 nakanishi:2 prevented:2 award:1 feasibility:1 prediction:2 regression:7 basic:3 controller:2 vision:1 represent:1 kernel:14 robotics:5 achieved:1 cps:24 affecting:2 addition:1 subject:2 tend:1 recording:3 flow:1 mod:3 jordan:1 feedforward:1 variety:3 affect:1 fit:2 independence:1 restrict:1 bandwidth:2 topology:1 drum:1 angeles:1 six:1 reuse:1 action:6 useful:1 amount:1 nonparametric:3 locally:6 simplest:1 http:2 canonical:9 dotted:3 track:3 per:3 discrete:16 four:2 drawn:1 clarity:1 asymptotically:2 angle:4 inverse:2 letter:4 topologically:1 place:1 almost:1 oscillation:3 dy:2 scaling:5 summarizes:1 jointed:1 hi:6 guaranteed:2 simplification:1 occur:1 bongo:1 encodes:1 aspect:1 speed:2 span:1 performing:1 flexion:1 ball:5 remain:1 smaller:1 character:8 wi:13 modification:3 leg:1 explained:1 invariant:1 indexing:1 equation:10 remains:1 previously:2 fed:2 end:1 serf:1 appropriate:2 generic:2 petsche:1 robustly:1 alternative:2 robustness:2 original:1 denotes:1 top:3 nakano:1 exploit:1 society:1 rsj:1 objective:1 ude:1 already:1 realized:1 added:2 alamitos:1 traditional:2 southern:1 lends:1 separate:1 atr:3 topic:1 trivial:1 toward:3 equip:1 reason:1 assuming:5 providing:1 demonstration:7 implementation:1 design:2 proper:1 policy:31 adjustable:1 perform:1 beat:4 extended:1 shoulder:1 head:1 communication:1 perturbation:6 arbitrary:2 nonlinearly:1 required:1 california:1 learned:14 temporary:1 established:1 able:3 suggested:3 usually:1 dynamical:11 below:1 pattern:8 max:1 memory:1 suitable:2 force:1 arm:6 representing:1 movie:1 technology:2 temporally:1 irrespective:1 carried:1 jun:2 literature:1 tangent:1 meter:1 asymptotic:1 law:1 interesting:3 allocation:1 proportional:1 analogy:2 foundation:1 humanoid:13 degree:2 editor:1 systematically:1 wearing:1 summary:1 surprisingly:1 repeat:1 soon:1 keeping:1 allow:2 institute:1 face:1 rhythmic:23 tracing:1 distributed:1 feedback:1 dimension:1 computes:1 made:1 reinforcement:5 atkeson:2 far:2 employing:1 cope:1 skill:6 keep:1 global:1 anchor:2 consuming:1 demo:4 imitation:5 disrupt:1 alternatively:1 continuous:6 table:2 learn:5 robust:1 ca:1 complex:5 domain:1 whole:1 profile:1 position:14 explicit:2 exponential:1 replay:2 third:1 schaal1:1 paused:1 specific:2 gating:1 hale:1 ci:9 illustrates:1 cartesian:2 suited:1 smoothly:2 hitting:1 doubling:1 monotonic:1 ch:2 extracted:1 ma:1 modulate:2 goal:14 formulated:4 oscillator:1 crl:1 change:5 typical:1 determined:1 wt:1 ijspeert:5 invariance:2 experimental:1 osu:1 support:1 modulated:3 constructive:1 evaluate:1 tested:1 scratch:3 |
1,253 | 2,141 | Global Versus Local Methods
in Nonlinear Dimensionality Reduction
Vin de Silva
Department of Mathematics,
Stanford University,
Stanford. CA 94305
[email protected]
Joshua B. Tenenbaum
Department of Brain and Cognitive Sciences,
Massachusetts Institute of Technology,
Cambridge. MA 02139
[email protected]
Abstract
Recently proposed algorithms for nonlinear dimensionality reduction fall
broadly into two categories which have different advantages and disadvantages: global (Isomap [1]), and local (Locally Linear Embedding [2],
Laplacian Eigenmaps [3]). We present two variants of Isomap which
combine the advantages of the global approach with what have previously been exclusive advantages of local methods: computational sparsity and the ability to invert conformal maps.
1 Introduction
In this paper we discuss the problem of nonlinear dimensionality reduction (NLDR): the
task of recovering meaningful low-dimensional structures hidden in high-dimensional data.
An example might be a set of pixel images of an individual?s face observed under different pose and lighting conditions; the task is to identify the underlying variables (pose angles, direction of light, etc.) given only the high-dimensional pixel image data. In many
cases of interest, the observed data are found to lie on an embedded submanifold of the
high-dimensional space. The degrees of freedom along this submanifold correspond to the
underlying variables. In this form, the NLDR problem is known as ?manifold learning?.
Classical techniques for manifold learning, such as principal components analysis (PCA)
or multidimensional scaling (MDS), are designed to operate when the submanifold is embedded linearly, or almost linearly, in the observation space. More generally there is a
wider class of techniques, involving iterative optimization procedures, by which unsatisfactory linear representations obtained by PCA or MDS may be ?improved? towards more
successful nonlinear representations of the data. These techniques include GTM [4], self
organising maps [5] and others [6,7]. However, such algorithms often fail when nonlinear
structure cannot simply be regarded as a perturbation from a linear approximation; as in
the Swiss roll of Figure 3. In such cases, iterative approaches tend to get stuck at locally
optimal solutions that may grossly misrepresent the true geometry of the situation.
Recently, several entirely new approaches have been devised to address this problem. These
methods combine the advantages of PCA and MDS?computational efficiency; few free
parameters; non-iterative global optimisation of a natural cost function?with the ability to
recover the intrinsic geometric structure of a broad class of nonlinear data manifolds.
These algorithms come in two flavors: local and global. Local approaches (LLE [2], Laplacian Eigenmaps [3]) attempt to preserve the local geometry of the data; essentially, they
seek to map nearby points on the manifold to nearby points in the low-dimensional representation. Global approaches (Isomap [1]) attempt to preserve geometry at all scales,
mapping nearby points on the manifold to nearby points in low-dimensional space, and
faraway points to faraway points.
The principal advantages of the global approach are that it tends to give a more faithful
representation of the data?s global structure, and that its metric-preserving properties are
better understood theoretically. The local approaches have two principal advantages: (1)
computational efficiency: they involve only sparse matrix computations which may yield
a polynomial speedup; (2) representational capacity: they may give useful results on a
broader range of manifolds, whose local geometry is close to Euclidean, but whose global
geometry may not be.
In this paper we show how the global geometric approach, as implemented in Isomap,
can be extended in both of these directions. The results are computational efficiency and
representational capacity equal to or in excess of existing local approaches (LLE, Laplacian
Eigenmaps), but with the greater stability and theoretical tractability of the global approach.
Conformal Isomap (or C-Isomap) is an extension of Isomap which is capable of learning the
structure of certain curved manifolds. This extension comes at the cost of making a uniform
sampling assumption about the data. Landmark Isomap (or L-Isomap) is a technique for
approximating a large global computation in Isomap by a much smaller set of calculations.
Most of the work focuses on a small subset of the data, called the landmark points.
The remainder of the paper is in two sections. In Section 2, we describe a perspective on
manifold learning in which C-Isomap appears as the natural generalisation of Isomap. In
Section 3 we derive L-Isomap from a landmark version of classical MDS.
2 Isomap for conformal embeddings
2.1 Manifold learning and geometric invariants
We can view the problem of manifold learning as an attempt to invert a generative model
for a set of observations. Let be a -dimensional domain contained in the Euclidean
, and let
be a smooth embedding, for some
. The object of
space
manifold learning is to recover and based on a given set
of observed data in
.
The observed data arise as follows. Hidden data
are generated randomly in , and are
then mapped by to become the observed data, so
.
The problem as stated is ill-posed: some restriction is needed on if we are to relate the
observed geometry of the data to the structure of the hidden variables
and itself. We
will discuss two possibilities. The first is that is an isometric embedding in the sense of
Riemannian geometry; so preserves infinitesmal lengths and angles. The second possibility is that is a conformal embedding; it preserves angles but not lengths. Equivalently,
at every point
there is a scalar
such that infinitesimal vectors at get magnified in length by a factor
. The class of conformal embeddings includes all isometric
embeddings as well as many other families of maps, including stereographic projections
such as the Mercator projection.
$#
%&'
)(
%*
!"
One approach to solving a manifold learning problem is to identify which aspects of the
geometry of are invariant under the mapping . For example, if is an isometric embedding then by definition infinitesimal distances are preserved. But more is true. The length
of a path in is defined by integrating the infinitesimal distance metric along the path.
The same is true in
, so preserves path lengths. If
are two points in , then
the shortest path between and lying inside is the same length as the shortest path
-
,+.-
'
-
between
and
along
. Thus geodesic distances are preserved. The conclusion is that is isometric with
, regarded as metric spaces under geodesic distance.
Isomap exploits this idea by constructing the geodesic metric for
approximately as a
matrix, using the observed data alone.
To solve the conformal embedding problem, we need to identify an observable geometric
invariant of conformal maps. Since conformal maps are locally isometric up to a scale
factor
, it is natural to try to estimate
at each point
in the observed data. By
rescaling, we can then restore the original metric structure of the data and proceed as in
Isomap. We can do this by noting that a conformal map rescales local volumes in by a
factor
. Hence if the
hidden data are sampled uniformly in , the local density of the
observed data will be
. It follows that the conformal factor
can be estimated
in terms of the observed local data density, provided that the original sampling is uniform.
C-Isomap implements a version of this idea which is independent of the dimension .
%*
%*
%&'
%* .
%*
This uniform sampling assumption may appear to be a severe restriction, but we believe
it reflects a necessary tradeoff in dealing with a larger class of maps. Moreover, as we
illustrate below, our algorithm appears in practice to be robust to moderate violations of
this assumption.
2.2 The Isomap and C-Isomap algorithms
There are three stages to Isomap [1]:
!
'
1. Determine a neighbourhood graph of the observed data
in a suitable way.
For example, might contain iff is one of the nearest neighbours of
(and vice versa). Alternatively, might contain the edge iff
,
for some .
,
,
2. Compute shortest paths in the graph for all pairs of data points. Each edge
in the graph is weighted by its Euclidean length
, or by some other useful
metric.
3. Apply MDS to the resulting shortest-path distance matrix
ding of the data in Euclidean space, approximating .
to find a new embed-
The premise is that local metric information (in this case, lengths of edges in the
neighbourhood graph) is regarded as a trustworthy guide to the local metric structure in the
original (latent) space. The shortest-paths computation then gives an estimate of the global
metric structure, which can be fed into MDS to produce the required embedding.
It is known that Step 2 converges on the true geodesic structure of the manifold given
sufficient data, and thus Isomap yields a faithful low-dimensional Euclidean embedding
whenever the function is an isometry. More precisely, we have (see [8]):
, with respect to a density
Theorem. Let be sampled from a bounded convex region in
function
.
Let
be
a
-smooth
isometric
embedding
of
that region in
. Given
, for a suitable choice of neighbourhood size parameter or , we have
+
)(
recovered distance
original distance
with probability at least
, provided that the sample size is sufficiently large.
formula is taken to hold for all pairs of points simultaneously.]
[The
C-Isomap is a simple variation on Isomap. Specifically, we use the -neighbours method
in Step 1, and replace Step 2 with the following:
2a. Compute shortest paths in the graph for
all pairs of data points. Each edge in
the graph is weighted by
is the mean distance
. Here
of to its nearest neighbours.
,
Using similar arguments to those in [8], one can prove a convergence theorem for CIsomap. The exact formula for the
weights
is not critical in the asymptotic analysis. The
point is that the rescaling factor
is an asymptotically accurate approximation
to the conformal scaling factor in the neighbourhood of and .
Theorem. Let be sampled uniformly from a bounded convex region
in
. Let be
a -smooth conformal embedding of that region in
. Given
, for a suitable
choice of neighbourhood size parameter , we have
recovered distance
original distance
with probability at least
, provided that the sample size is sufficiently large.
+
(
It is possible but unpleasant to find explicit lower bounds for the sample size. Qualitatively, we expect to require a larger sample size for C-Isomap since it depends on two
approximations?local data density and geodesic distance?rather than one. In the special
case where the conformal embedding is actually an isometry, it is therefore preferable to
use Isomap rather than C-Isomap. This is borne out in practice.
2.3 Examples
We ran C-Isomap, Isomap, MDS and LLE on three ?fishbowl? examples with different data
distributions, as well as a more realistic simulated data set. We refer to Figure 1.
Fishbowls: These three datasets differ only in the probability density used to generate
the points. For the conformal fishbowl (column 1), 2000 points were generated randomly
and then projected stereographically (hence conformally
uniformly in a circular disk
mapped) onto a sphere. Note the high concentration of points near the rim. There is no
metrically faithful way of embedding a curved fishbowl inside a Euclidean plane, so classical MDS and Isomap cannot succeed. As predicted, C-Isomap does recover the original
disk structure of (as does LLE). Contrast with the uniform fishbowl (column 2), with data
points sampled using a uniform measure on the fishbowl itself. In this situation C-Isomap
behaves like Isomap, since the rescaling factor is approximately constant; hence it is unable
to find a topologically faithful 2-dimensional representation. The offset fishbowl (column 3)
is a perturbed version of the conformal fishbowl; points are sampled in using a shallow
Gaussian offset from center, then stereographically projected onto a sphere. Although the
theoretical conditions for perfect recovery are not met, C-Isomap is robust enough to find a
topologically correct embedding. LLE, in contrast, produces topological errors and metric
distortion in both cases where the data are not uniformly sampled in (columns 2 and 3).
Face images: Artificial images of a face were rendered as
pixel images and rasterized into 16384-dimensional vectors. The images varied randomly and independently
in two parameters: left-right pose angle and distance from camera . There is a natural
family of conformal transformations for this
if we ignore perspective distor data manifold,
, for
, which has the effect of shrinking
tions in the closest images: namely
or magnifying the apparent size of images by a constant factor. Sampling uniformly in
and in gives a data set approximately satisfying the required conditions for C-Isomap.
We generated 2000 face images in this way, spanning the range indicated by Figure 2. All
four algorithms returned a two-dimensional embedding of the data. As expected, C-Isomap
returns the cleanest embedding, separating the two degrees of freedom reliably along the
horizontal and vertical axes. Isomap returns an embedding which narrows predictably as
the face gets further away. The LLE embedding is highly distorted.
(
conformal fishbowl
uniform fishbowl
offset fishbowl
face images
MDS
MDS
MDS
MDS
Isomap: k = 15
Isomap: k = 15
Isomap: k = 15
Isomap: k = 15
C?Isomap: k = 15
C?Isomap: k = 15
C?Isomap: k = 15
C?Isomap: k = 15
LLE: k = 15
LLE: k = 15
LLE: k = 15
LLE: k = 15
Figure 1: Four dimensionality reduction algorithms (MDS, Isomap, C-Isomap, and LLE)
are applied to three versions of a toy ?fishbowl? dataset, and to a more complex data manifold of face images.
Figure 2: A set of 2000 face images were randomly generated, varying independently in
two parameters: distance and left-right pose. The four extreme cases are shown.
3 Isomap with landmark points
The Isomap algorithm has two computational bottlenecks. The first is calculating the
shortest-paths distance matrix
. Using Floyd?s algorithm this is
; this can be
improved to
by implementing Dijkstra?s algorithm with Fibonacci heaps
( is the neighbourhood size). The second bottleneck is the MDS eigenvalue calculation,
which involves a full
matrix and has complexity
. In contrast, the eigenvalue
computations in LLE and Laplacian Eigenmaps are sparse (hence considerably cheaper).
L-Isomap addresses both of these inefficiencies at once. We designate of the data points to
be landmark points, where
. Instead of computing
, we compute the
matrix
of distances from each data point to the landmark points only. Using a new procedure LMDS (Landmark MDS), we find a Euclidean embedding of the data using
instead of
. This leads to an enormous savings when is much less than , since
can be computed using Dijkstra in
time, and LMDS runs in
.
LMDS is feasible precisely because we expect the data to have a low-dimensional embedding. The first step is to apply classical MDS to the landmark points only, embedding them
faithfully in . Each remaining point can now be located in
by using its known distances from the landmark points as constraints. This is analogous to the Global Positioning
System technique of using
a finite number of distance readings to identify a geographic
location. If
and the landmarks are in general position, then there are enough
constraints to locate uniquely. The landmark points
may be chosen randomly, with
taken to be sufficiently larger than the minimum
to ensure stability.
3.1 The Landmark MDS algorithm
LMDS begins by applying classical MDS [9,10] to the landmarks-only distance matrix
. We recall
the procedure. The first step is to construct an ?inner-product? matrix
; here
is the matrix of squared
distances and
is the ?centering? matrix
. Next find the eigenvalues
and eigenvectors
of
defined by the formula
. Write for the positive eigenvalues (labelled so that
), and for
the corresponding eigenvectors
(written as column vectors); non-positive eigenvalues are
ignored. Then for
the required optimal -dimensional embedding vectors are given
as the columns of the matrix:
..
.
!"
&
"
# " %$
0/
('
*))+) ,
.-
233 5 6'87 .:'9 ;=<<
3 5 7 . -- 9 <
1
4 5 7 >
.- 9
The embedded data are automatically mean-centered with principal components aligned
/
with the axes, most significant first. If ? has no negative eigenvalues, then the dimensional embedding is perfect; otherwise there is no exact Euclidean embedding.
@
"A
The second stage of LMDS is to embed the remaining points in
. Let
denote the
column vector of squared distances between a data point and the landmark points. The
embedding vector is related linearly to
by the formula:
-
where
D
is the column mean of
A
C
1
B
-
D "A&
1
1
and 23 B is the
pseudoinverse transpose of
33 .:- '9 55 6 ' ; <<<
1CB .:- 9 .
4 ..5 >
.- 9
:
Original points
L?Isomap: k=8
20 landmarks
L?Isomap: k=8
10 landmarks
L?Isomap: k=8
4 landmarks
L?Isomap: k=8
3 landmarks
Swiss roll embedding
LLE: k=18
LLE: k=14
LLE: k=10
LLE: k=6
Figure 3: L-Isomap is stable over a wide range of values for the sparseness parameter
(the number of landmarks). Results from LLE are shown for comparision.
The final (optional) stage is to use PCA to realign the data with the coordinate axes. A full
discussion of LMDS will appear in [11]. We note two results here:
1. If is a landmark point, then the embedding given by LMDS is consistent with
the original MDS embedding.
2. If the distance matrix
can be represented exactly by a Euclidean config, and if the landmarks are chosen so that their affine span in that
uration in
configuration is -dimensional (i.e. in general position), then LMDS will recover
the configuration exactly, up to rotation and translation.
A good way to satisfy the affine span condition is to pick
landmarks randomly, plus
a few extra for stability. This is important for Isomap, where the distances are inherently
slightly noisy. The robustness of LMDS to noise depends on the matrix norm
. If
is very small, then all the landmarks lie close to a hyperplane and LMDS
performs poorly with noisy data. In practice, choosing a few extra landmark points gives
satisfactory results.
5
1 B
3.2 Example
Figure 3, shows the results of testing L-Isomap on a Swiss roll data set. 2000 points were
generated uniformly in a rectangle (top left) and mapped into a Swiss roll configuration
in
. Ordinary Isomap recovers the rectangular structure correctly
provided that the
neighbourhood parameter is not too large (in this case
works). The tests show
that this peformance is not significantly degraded when L-Isomap is used. For each , we
chose landmark points at random; even down to 4 landmarks the embedding closely approximates the (non-landmark) Isomap embedding. The configuration of three landmarks
was chosen especially to illustrate the affine distortion that may arise if the landmarks lie
close to a subspace (in this case, a line). For three landmarks chosen at random, results are
generally much better.
In contrast, LLE is unstable under changes in its sparseness parameter (neighbourhood
size). To be fair, is principally a topological parameter and only incidentally a sparseness
parameter for LLE. In L-Isomap, these two roles are separately fulfilled by and .
4 Conclusion
Local approaches to nonlinear dimensionality reduction such as LLE or Laplacian Eigenmaps have two principal advantages over a global approach such as Isomap: they tolerate a
certain amount of curvature and they lead naturally to a sparse eigenvalue problem. However, neither curvature tolerance nor computational sparsity are explicitly part of the formulation of the local approaches; these features emerge as byproducts of the goal of trying
to preserve only the data?s local geometric structure. Because they are not explicit goals but
only convenient byproducts, they are not in fact reliable features of the local approach. The
conformal invariance of LLE can fail in sometimes surprising ways, and the computational
sparsity is not tunable independently of the topological sparsity of the manifold. In contrast, we have presented two extensions to Isomap that are explicitly designed to remove
a well-characterized form of curvature and to exploit the computational sparsity intrinsic
to low-dimensional manifolds. Both extensions are amenable to algorithmic analysis, with
provable conditions under which they return accurate results; and they have been tested
successfully on challenging data sets.
Acknowledgments
This work was supported in part by NSF grant DMS-0101364, and grants from Schlumberger, MERL and the DARPA Human ID program. The authors wish to thank Thomas
Vetter for providing the range and texture maps for the synthetic face; and Lauren Schmidt
for her help in rendering the actual images using Curious Labs? ?Poser? software.
References
[1] Tenenbaum, J.B., de Silva, V. & Langford, J.C (2000) A global geometric framework for nonlinear
dimensionality reduction. Science 290: 2319?2323.
[2] Roweis, S. & Saul, L. (2000) Nonlinear dimensionality reduction by locally linear embedding.
Science 290: 2323?2326.
[3] Belkin, M. & Niyogi, P. (2002) Laplacian eigenmaps and spectral techniques for embedding and
clustering. In T.G. Dietterich, S. Becker and Z. Ghahramani (eds.), Advances in Neural Information
Processing Systems 14. MIT Press.
[4] Bishop, C., Svensen, M. & Williams, C. (1998) GTM: The generative topographic mapping.
Neural Computation 10(1).
[5] Kohonen, T. (1984) Self Organisation and Associative Memory. Springer-Verlag, Berlin.
[6] Bregler, C. & Omohundro, S.M. (1995) Nonlinear image interpolation using manifold learning. In
G. Tesauro, D.S. Touretzky & T.K. Leen (eds.), Advances in Neural Information Processing Systems
7: 973?980. MIT Press.
[7] DeMers, D. & Cottrell, G. (1993) Non-linear dimensionality reduction In S. Hanson, J. Cowan &
L. Giles (eds.), Advances in Neural Information Processing Systems 5: 580?590. Morgan-Kaufmann.
[8] Bernstein, M., de Silva, V., Langford, J.C. & Tenenbaum, J.B. (December 2000) Graph approximations to geodesics on embedded manifolds. Preprint may be downloaded at the URL:
http://isomap.stanford.edu/BdSLT.pdf
[9] Torgerson, W.S. (1958) Theory and Methods of Scaling. Wiley, New York.
[10] Cox, T.F. & Cox M.A.A. (1994) Multidimensional Scaling. Chapman & Hall, London.
[11] de Silva, V. & Tenenbaum, J.B. (in preparation) Sparse multidimensional scaling using landmark
points.
| 2141 |@word cox:2 version:4 polynomial:1 norm:1 disk:2 seek:1 pick:1 reduction:8 inefficiency:1 configuration:4 poser:1 existing:1 recovered:2 trustworthy:1 surprising:1 written:1 cottrell:1 realistic:1 remove:1 designed:2 alone:1 generative:2 plane:1 math:1 location:1 organising:1 along:4 become:1 prove:1 combine:2 inside:2 theoretically:1 expected:1 nor:1 brain:1 automatically:1 actual:1 provided:4 begin:1 underlying:2 moreover:1 bounded:2 what:1 magnified:1 transformation:1 every:1 multidimensional:3 preferable:1 fishbowl:12 exactly:2 grant:2 appear:2 positive:2 understood:1 local:19 tends:1 id:1 path:10 interpolation:1 approximately:3 mercator:1 might:3 plus:1 chose:1 challenging:1 range:4 faithful:4 stereographic:1 camera:1 testing:1 acknowledgment:1 practice:3 implement:1 swiss:4 procedure:3 significantly:1 projection:2 convenient:1 integrating:1 vetter:1 get:3 cannot:2 close:3 onto:2 applying:1 restriction:2 map:9 center:1 williams:1 independently:3 convex:2 rectangular:1 recovery:1 regarded:3 embedding:31 stability:3 variation:1 coordinate:1 analogous:1 exact:2 satisfying:1 located:1 observed:11 role:1 preprint:1 ding:1 region:4 ran:1 complexity:1 geodesic:6 solving:1 lmds:10 efficiency:3 darpa:1 gtm:2 represented:1 describe:1 london:1 artificial:1 choosing:1 whose:2 apparent:1 stanford:4 posed:1 solve:1 larger:3 distortion:2 otherwise:1 ability:2 niyogi:1 topographic:1 itself:2 noisy:2 final:1 associative:1 advantage:7 eigenvalue:7 product:1 remainder:1 kohonen:1 aligned:1 iff:2 poorly:1 representational:2 roweis:1 lauren:1 convergence:1 produce:2 incidentally:1 perfect:2 converges:1 object:1 wider:1 derive:1 illustrate:2 tions:1 svensen:1 pose:4 help:1 nearest:2 recovering:1 implemented:1 predicted:1 come:2 involves:1 met:1 differ:1 direction:2 closely:1 correct:1 centered:1 human:1 implementing:1 require:1 premise:1 designate:1 bregler:1 extension:4 hold:1 lying:1 sufficiently:3 hall:1 cb:1 mapping:3 algorithmic:1 heap:1 vice:1 faithfully:1 successfully:1 reflects:1 weighted:2 mit:3 gaussian:1 rather:2 varying:1 broader:1 ax:3 focus:1 unsatisfactory:1 rasterized:1 contrast:5 sense:1 hidden:4 her:1 pixel:3 ill:1 special:1 equal:1 once:1 saving:1 construct:1 sampling:4 chapman:1 broad:1 others:1 few:3 belkin:1 randomly:6 neighbour:3 preserve:6 simultaneously:1 individual:1 cheaper:1 geometry:8 attempt:3 freedom:2 schlumberger:1 interest:1 possibility:2 circular:1 highly:1 severe:1 violation:1 extreme:1 light:1 amenable:1 accurate:2 edge:4 capable:1 byproduct:2 necessary:1 euclidean:9 theoretical:2 merl:1 column:8 giles:1 disadvantage:1 ordinary:1 cost:2 tractability:1 subset:1 uniform:6 submanifold:3 successful:1 eigenmaps:6 too:1 perturbed:1 considerably:1 synthetic:1 density:5 squared:2 borne:1 cognitive:1 rescaling:3 return:3 toy:1 de:4 includes:1 rescales:1 satisfy:1 explicitly:2 depends:2 view:1 try:1 lab:1 recover:4 vin:1 degraded:1 roll:4 kaufmann:1 correspond:1 identify:4 yield:2 lighting:1 touretzky:1 whenever:1 ed:3 definition:1 infinitesimal:3 centering:1 grossly:1 dm:1 naturally:1 riemannian:1 recovers:1 sampled:6 demers:1 dataset:1 tunable:1 massachusetts:1 recall:1 dimensionality:8 rim:1 actually:1 appears:2 tolerate:1 isometric:6 improved:2 formulation:1 leen:1 stage:3 langford:2 horizontal:1 nonlinear:10 indicated:1 believe:1 effect:1 dietterich:1 contain:2 geographic:1 true:4 isomap:63 hence:4 jbt:1 satisfactory:1 floyd:1 self:2 uniquely:1 trying:1 pdf:1 omohundro:1 fibonacci:1 performs:1 silva:5 image:14 recently:2 rotation:1 behaves:1 volume:1 approximates:1 refer:1 significant:1 cambridge:1 versa:1 ai:1 mathematics:1 stable:1 etc:1 curvature:3 closest:1 isometry:2 perspective:2 moderate:1 tesauro:1 certain:2 verlag:1 joshua:1 preserving:1 minimum:1 greater:1 morgan:1 determine:1 shortest:7 full:2 smooth:3 positioning:1 characterized:1 calculation:2 sphere:2 devised:1 laplacian:6 variant:1 involving:1 optimisation:1 essentially:1 metric:10 sometimes:1 invert:2 preserved:2 separately:1 extra:2 operate:1 tend:1 cowan:1 december:1 misrepresent:1 curious:1 near:1 noting:1 bernstein:1 config:1 embeddings:3 enough:2 peformance:1 rendering:1 inner:1 idea:2 tradeoff:1 bottleneck:2 pca:4 url:1 becker:1 returned:1 proceed:1 york:1 ignored:1 generally:2 useful:2 involve:1 eigenvectors:2 amount:1 tenenbaum:4 locally:4 category:1 generate:1 http:1 nsf:1 estimated:1 fulfilled:1 correctly:1 torgerson:1 broadly:1 write:1 four:3 enormous:1 neither:1 rectangle:1 graph:7 asymptotically:1 run:1 angle:4 distorted:1 topologically:2 almost:1 family:2 scaling:5 entirely:1 bound:1 topological:3 comparision:1 precisely:2 constraint:2 software:1 nearby:4 aspect:1 argument:1 span:2 rendered:1 speedup:1 department:2 smaller:1 slightly:1 shallow:1 making:1 invariant:3 principally:1 taken:2 previously:1 discus:2 fail:2 nldr:2 needed:1 fed:1 conformal:18 apply:2 away:1 spectral:1 neighbourhood:8 schmidt:1 robustness:1 original:8 thomas:1 top:1 remaining:2 include:1 ensure:1 clustering:1 calculating:1 exploit:2 ghahramani:1 especially:1 approximating:2 classical:5 concentration:1 exclusive:1 md:19 subspace:1 distance:22 unable:1 mapped:3 simulated:1 capacity:2 landmark:31 separating:1 thank:1 berlin:1 manifold:19 unstable:1 spanning:1 provable:1 length:8 providing:1 equivalently:1 relate:1 stated:1 negative:1 reliably:1 vertical:1 observation:2 datasets:1 finite:1 curved:2 dijkstra:2 optional:1 situation:2 extended:1 locate:1 perturbation:1 varied:1 pair:3 required:3 namely:1 hanson:1 narrow:1 address:2 below:1 sparsity:5 reading:1 program:1 including:1 reliable:1 memory:1 suitable:3 critical:1 natural:4 restore:1 technology:1 geometric:6 asymptotic:1 embedded:4 expect:2 versus:1 downloaded:1 degree:2 affine:3 sufficient:1 consistent:1 realign:1 translation:1 supported:1 free:1 transpose:1 guide:1 lle:21 conformally:1 institute:1 fall:1 wide:1 face:9 saul:1 magnifying:1 emerge:1 sparse:4 tolerance:1 dimension:1 stuck:1 qualitatively:1 author:1 projected:2 excess:1 observable:1 ignore:1 dealing:1 global:16 pseudoinverse:1 predictably:1 alternatively:1 iterative:3 latent:1 robust:2 ca:1 inherently:1 complex:1 constructing:1 domain:1 cleanest:1 linearly:3 noise:1 arise:2 fair:1 wiley:1 shrinking:1 position:2 explicit:2 wish:1 lie:3 theorem:3 formula:4 embed:2 down:1 bishop:1 offset:3 organisation:1 intrinsic:2 metrically:1 texture:1 sparseness:3 flavor:1 simply:1 faraway:2 contained:1 scalar:1 springer:1 ma:1 succeed:1 goal:2 towards:1 labelled:1 replace:1 feasible:1 change:1 generalisation:1 specifically:1 uniformly:6 hyperplane:1 principal:5 called:1 invariance:1 meaningful:1 unpleasant:1 preparation:1 tested:1 |
1,254 | 2,142 | Dopamine Induced Bistability Enhances
Signal Processing in Spiny Neurons
Aaron J. Gruber l ,2, Sara A. Solla2,3, and James C. Houk 2,l
Departments of Biomedical Engineeringl, Physiology2, and Physics and Astronomy3
Northwestern University, Chicago, IL 60201
{ a-gruberl, solla, j-houk }@northwestern.edu
Abstract
Single unit activity in the striatum of awake monkeys shows a
marked dependence on the expected reward that a behavior will
elicit. We present a computational model of spiny neurons, the
principal neurons of the striatum, to assess the hypothesis that direct neuromodulatory effects of dopamine through the activation
of D 1 receptors mediate the reward dependency of spiny neuron
activity. Dopamine release results in the amplification of key ion
currents, leading to the emergence of bistability, which not only
modulates the peak firing rate but also introduces a temporal and
state dependence of the model's response, thus improving the detectability of temporally correlated inputs.
1
Introduction
The classic notion of the basal ganglia as being involved in purely motor processing
has expanded over the years to include sensory and cognitive functions. A surprising new finding is that much of this activity shows a motivational component.
For instance, striatal activity related to visual stimuli is dependent on the type of
reinforcement (primary vs secondary) that a behavior will elicit [1]. Task-related
activity can be enhanced or suppressed when a reward is anticipated for correct
performance, relative to activity when no reward is expected. Although the origin
of this reward dependence has not been experimentally verified, dopamine modulation is likely to playa role. Spiny neurons in the striatum, the input to the basal
ganglia, receive a prominent neuromodulatory input from dopamine neurons in the
substantia nigra pars compacta. These dopamine neurons discharge in a rewarddependent manner [2]; they respond to the delivery of unexpected rewards and to
sensory cues that reliably precede the delivery of expected rewards.
Activation of dopamine receptors alters the response characteristics of spiny neurons
by modulating the properties of voltage-gated ion channels, as opposed to simple
excitatory or inhibitory effects [3]. Activation of the D1 type dopamine receptor
alone can either enhance or suppress neural responses depending on the prior state
of the spiny neuron [4]. Here, we use a computational approach to assess the
hypothesis that the modulation of specific ion channels through the activation of
D1 receptors is sufficient to explain both the enhanced and suppressed single unit
responses of medium spiny neurons to reward-predicting stimuli.
We have constructed a biophysically grounded model of a spiny neuron and used it to
investigate whether dopamine neuromodulation accounts for the observed rewarddependence of striatal single-unit responses to visual targets in the memory guided
saccade task described by [1]. These authors used an asymmetric reward schedule and compared the response to a given target in rewarded as opposed to unrewarded cases. They report a substantial reward-dependent difference; the majority
of these neurons showed a reward-related enhancement of the intensity and duration
of discharge, and a smaller number exhibited a reward-related depression. The authors speculated that D1 receptor activation might account for enhanced responses,
whereas D2 receptor activation might explain the depressed responses. The model
presented here demonstrates that neuromodulatory actions of dopamine through
D1 receptors suffice to account for both effects, with interesting consequences for
information processing.
2
Model description
The membrane properties of the model neuron result from an accurate representation of a minimal set of currents needed to reproduce the characteristic behavior of
spiny neurons. In low dopamine conditions, these cells exhibit quasi two-state behavior; they spend most of their time either in a hyperpolarized 'down' state around
-85 mV, or in a depolarized 'up' state around -55 mV [5]. This bimodal character
of the response to cortical input is attributed to a combination of inward rectifying (IRK) and outward rectifying (ORK) potassium currents [5]. IRK contributes
a small outward current at hyperpolarized membrane potentials, thus providing
resistance to depolarization and stabilizing the down state. ORK is a major hyperpolarizing current that becomes activated at depolarized potentials and opposes
the depolarizing influences of excitatory synaptic and inward ionic currents; it is
their balance that determines the membrane potential of the up state. In addition
to IRK and ORK currents, the model incorporates the L-type calcium (L-Ca) current that starts to provide an inward current at subthreshold membrane potentials,
thus determining the voltage range of the up state. This current has the ability to
increase the firing rate of spiny neurons and is critical to the enhancement of spiny
neuron responses in the presence of D1 agonists [4].
Our goal is to design a model that provides a consistent description of membrane
properties in the 100 - 1000 ms time range. This is the characteristic range of
duration for up and down state episodes; it also spans the time course of short
term modulatory effects of dopamine. The model is constructed according to the
principle of separation of time scales: processes that operate in the 100-1000 ms
range are modeled as accurately as possible, those that vary on a much shorter time
scale are assumed to instantaneously achieve their steady-state values, and those
that occur over longer time scales, such as slow inactivation, are assumed constant.
Thus, the model does not incorporate currents which inactivate on a short time
scale, and cannot provide a good description of rapid events such as the transitions
between up and down states or the generation of action potentials.
The membrane of a spiny neuron is modeled here as a single compartment with
steady-state voltage-gated ion currents. A first order differential equation relates
the temporal change in membrane potential (Vm ) to the membrane currents (Ii),
(1)
The right hand side of the equation includes active ionic, leakage, and synaptic
currents. The multiplicative factor 'Y models the modulatory effects of D1 receptor
activation by dopamine, to be described in more detail later. Ionic currents are
modeled using a standard formulation; the parameters are as reported in the biophysical literature, except for adjustments that compensate for specific experimental
conditions so as to more closely match in vivo realizations.
All currents except for L-Ca are modeled by the product of a voltage gated conductance and a linear driving force , Ii = gi (Vm - E i ), where Ei is the reversal potential
of ion species i and gi is the corresponding conductance. The leakage conductance
is constant; the conductances for IRK and ORK are voltage gated, gi = ?hLi (Vm ),
where 9i is the maximum conductance and Li (Vm ) is a logistic function of the
membrane potential. Calcium currents are not well represented by a linear driving
force model; extremely low intracellular calcium concentrations result in a nonlinear
driving force well accounted for by the Goldman-Hodgkin-Katz equation [6],
h -C a
=
PL -C aLL -C a (Vm )
2) ([Ca] [Ca]oe _
(z 2VmF
RT
1 _ e-?r
i -
zV",F
)
RT
'
(2)
where FL -C a is the maximum permeability. The resulting ionic currents are shown
in Fig 1A.
The synaptic current is modeled as the product of a conductance and a linear
driving force, 18 = g8(Vm - E 8), with E8 = O. The synaptic conductance includes
two types of cortical input: a phasic sensory-related component gp, and a tonic
context-related component gt, which are added to determine the total synaptic
input: g8 = ~(gp + gt). The factor ~ is a random variable that simulates the noisy
character of synaptic input.
Dopamine modulates the properties of ion currents though the activation of specific
receptors. Agonists for the D1 type receptor enhance the IRK and L-Ca currents
observed in spiny neurons [7, 8]. This effect is modeled by the factor 'Y in Eq 1.
An upper bound of'Y = 1.4 is derived from physiological experiments [7, 8]. The
lower bound at 'Y = 1.0 corresponds to low dopamine levels; this is the experimental
condition in which the ion currents have been characterized.
3
Static and dynamic properties
Stationary solutions to Eq 1 correspond to equilibrium values of the membrane
potential Vm consistent with specific values of the dopamine controlled conductance
gain parameter 'Y and the total synaptic conductance g8; fluctuations of g8 around
its mean value are ignored in this section: the noise parameter is set to ~ = 1.
Stationary solutions satisfy dVm/dt = 0; it follows from Eq 1 that they result from
(3)
Intersections between a curve representing the total ionic current (left hand side of
Eq 3) as a function of Vm and a straight line representing the negative of the synaptic
current (right hand side of Eq 3) determine the stationary values of the membrane
potential. Solutions to Eq 3 can be followed as a function of g8 for fixed 'Y by varying
the slope of the straight line. For 'Y = 1 there is only one such intersection for any
value of g8. At low dopamine levels, Vm is a single-valued monotonically increasing
function of g8, shown in Fig 1B (dotted line). This operational curve describes a
A
B
-30
2
>-50
N
E
()
::c 0 +--1----=::::::__
.sE
>
.6
-70
-90.j....i~=-.:::::;:...---~
-80
-60
Vm(mV)
o
10
9s (IlS/cm2)
20
Figure 1: Model characterization in low (-y = 1.0, dotted lines) and high (-y = 1.4,
solid lines) dopamine conditions. (A) Voltage-gated ion currents. (B) Operational
curves: stationary solutions to Eq 1.
gradual, smooth transition from hyperpolarized values of Vm corresponding to the
down state to depolarized values of Vm corresponding to the up state. At high
dopamine levels (-y = 1.4), the membrane potential is a single-valued monotonically
increasing function of the synaptic conductance for either g8 < 9.74 JLS/cm 2 or
g8 > 14.17 JLS/cm 2 . In the intermediate regime 9.74 JLS/cm 2 < g8 < 14.17 JLS/cm 2 ,
there are three solutions to Eq 3 for each value of g8. The resulting operational
curve, shown Fig 1B (solid line), consists of three branches: two stable and one
unstable. The two stable branches (dark solid lines) correspond to a hyperpolarized
down state (lower branch) and a depolarized up state (upper branch). The unstable
branch (solid gray line) corresponds to intermediate values of Vm that are not
spontaneously sustainable.
Bistability arises through a saddle node bifurcation with increasing 'Y and has a
drastic effect on the response properties of the model neuron in high dopamine
conditions. Consider an experiment in which 'Y is fixed at 1.4 and g8 changes slowly
so as to allow Vm to follow its equilibrium value on the operational curve for 'Y = 1.4
(see Fig 1B). As g8 increases, the hyperpolarized down state follows the lower stable
branch. As g8 reaches 14.17 JLS/cm 2 , the synaptic current suddenly overcomes the
mostly IRK hyperpolarizing current, and Vm depolarizes abruptly to reach an up
state stabilized by the activation of the hyperpolarizing aRK current. This is
the down to up (D-+U) state transition. As g8 is increased further, the up state
follows the upper stable branch, with a small amount of additional depolarization.
If g8 is now decreased, the depolarized up state follows the stable upper branch
in the downward direction. It is the inward L-Ca current which counteracts the
hyperpolarizing effect of the aRK current and stabilizes the up state until g8 reaches
9.74 JLS/cm 2 , where a net hyperpolarizing ionic current overtakes the system and
Vm hyperpolarizes abruptly to the down state. This is the up to down (U-+D) state
transition.
The emergence of bistability in high dopamine conditions results in a prominent
hysteresis effect. The state of the model, as described by the value of Vm , depends
not only on the current values of 'Y and g8' but also on the particular trajectory
followed by these parameters to reach their current values. The appearance of
bistability gives a well defined meaning to the notion of a down state and an up state:
in this case there is a gap between the two stable branches, while in low dopamine
conditions the transition is smooth, with no clear separation between states. We
generically refer to hyperpolarized potentials as the down state and depolarized
potentials as the up state, for consistency with the electrophysiological terminology.
A
Unrewarded Trial
Rewarded Trial
-30
-9p
-9p
-50
g> VI
:-Da
+/ .........
(
.. ~ .. '
E
> -70
;,...
.......;.;;.::.:.~p+Da
::'
.....
...
?????[j-+?~9p~
-90
0
B
Unrewarded Trial
Rewarded Trial
-30
-9p
>-50
VI
. .. ...
~ ... l
g
~
,
E
> -70
I'
.... ...
-90
0
.....
......
... ...
....
8
+9p
+Da
:
6.:+~p:
95 (J.tS/cm2)
......
,"
-Da
.....
16
0
8
95
16
95 (J.tS/cm2)
Figure 2: Response to a sensory related phasic input in rewarded and unrewarded
trials. (A) gt + gp > gD-+U? (B) gt + gp < g;.
An important feature of the model is that operational curves for all values of, intersect at a unique point, indicated by a circle in Fig 1B, for which V;' = - 55.1 m V
and g; = 13.2 J-tS / cm 2 . The appearance of this critical point is due to a perfect cancellation between the IRK and the L-Ca currents; it arises as a solut ion
to the equation I IRK + h-ca = O. When this condition is satisfied, solutions
to Eq 3 become independent of,. The existence of a critical point at a slightly
more depolarized membrane potential than the firing threshold at VI = - 58 m V
is an important aspect of our model; it plays a role in the mechanism that allows
dopamine to either enhance or depress the response of the model spiny neuron.
The dynamical evolution of Vm due to changes in both g8 and, follows from Eq 1.
Consider a scenario in which a tonic input gt maintains Vm below VI; the response to
an additional phasic input gp sufficient to drive Vm above VI depends on whether it
is associated with expected reward and thus triggers dopamine release. The response
of the model neuron depends on the combined synaptic input g8 in a manner that is
critically dependent on the expectation of reward. We consider two cases: whether
g8 exceeds gD-+U (Fig 2A) or remains below g; (Fig 2B).
If the phasic input is not associated with reward, the dopamine level does not increase (left panels in Fig 2). The square on the operational curve for , = 1 (dotted
line) indicates the equilibrium state corresponding to gt. A rapid increase from
g8 = gt to g8 = gt + gp (rightward solid arrow) is followed by an increase in Vm
towards its equilibrium value (upward dotted arrow). When the phasic input is
removed (leftward solid arrow), Vm decreases to its initial equilibrium value (down-
9D-U
9;
enhanced
amplitude
9t
N
E
c75
6
enhanced
amplitude
and
7.5
d)
No Response
O-l-----~-~___>,~"t_,
o
Figure 3: Modulation
of response in high
dopamine relative to
low dopamine conditions
as a function of the
strength of phasic and
tonic inputs.
ward dotted arrow). In unrewarded trials, the only difference between a larger and
a smaller phasic input is that the former results in a more depolarized membrane
potential and thus a higher firing rate. The firing activity, which ceases when the
phasic input disappears, encodes for the strength of the sensory-related stimulus.
Rewarded trials (right panels in Fig 2) elicit qualitatively different responses. The
phasic input is the conditioned stimulus that triggers dopamine release in the striatum, and the operational curve switches from the '"Y = 1 (dotted) curve to the
bistable '"Y = 1.4 (solid) curve. The consequences of this switch depend on the
strength of the phasic input. If g8 exceeds the value for the D-+ U transition (Fig
2A) , Vm depolarizes towards the upper branch of the bistable operational curve.
This additional depolarization results in a noticeably higher firing rate than the
one elicited by the same input in an unrewarded trial (Fig 2A, left panel). When
the phasic input is removed, the unit hyperpolarizes slightly as it reaches the upper
branch of the bistable operational curve. If gt exceeds gU--+D, the unit remains in
the up state until '"Y decreases towards its baseline level. If this condition is met
in a rewarded trial, the response is not only larger in amplitude but also longer in
duration. In contrast to these enhancements, if g8 is not sufficient to exceed g; (Fig
2B), Vm hyperpolarizes towards the lower branch of the bistable operational curve.
The unit remains in the down state until '"Y decreases towards its baseline level. In
this type of rewarded trial, dopamine suppresses the response of the unit.
The analysis presented above provides an explanatory mechanism for the observation of either enhanced or suppressed spiny neuron activity in the presence of
dopamine. It is the strength of the total synaptic input that selects between these
two effects; the generic features of their differentiation are summarized in Fig 3.
Enhancement occurs whenever the condition g8 > gD--+U is met, while activity is
suppressed if g8 < g;. The separatrix between enhancement and suppression always lies in a narrow band limited by g8 = gD--+U and g8 = g;. Its precise location
will depend on the details of the temporal evolution of '"Y as it rises and returns to
baseline. But whatever the shape of '"Y(t) might be, there will be a range of values
of g8 for which activity is suppressed, and a range of values of g8 for which activity
is enhanced.
4
Information processing
Dopamine induced bistability improves the ability of the model spiny neuron to detect time correlated sensory-related inputs relative to a context-related background.
To illustrate this effect, consider g8 = ~(gt + gp) as a random variable. The multiplicative noise is Gaussian, with <~>= 1 and <e >= 1.038. The total probability
density function (PDF) shown in Fig 4A for gt = 9.2 J-LS/cm 2 consists of two PDFs
corresponding to gp = 0 (left; black line) and gp = 5.8 J-LS/cm 2 (right; grey line).
These two values of gp occur with equal prior probability; time correlations are
introduced through a repeat probability Pr of retaining the current value of gp in
the subsequent time step. The total PDF shown in Fig 4A does not depend on the
value of Pr. Performance at the task of detecting the sensory-related input (gp -::f- 0)
is limited by the overlap of the corresponding PDFs [9]; optimal separation of the
two PDFs in Fig 4A results in a Bayesian error of 10.46%.
c
B
A
D
o
-60
Vm (mV)
Vm (mV)
-30
Vm (mV)
Figure 4: Probability density functions for (A) synaptic input, (B) membrane potential at "( = 1, (C) membrane potential at "( = 1.4 for un correlated inputs (Pr = 0.5) ,
and (D) membrane potential at "( = 1.4 for correlated inputs (Pr = 0.975).
The transformation of g8 into Vm through the "( = 1 operational curve results in
the PDFs shown in Fig 4B; here again, the total PDF does not depend on Pr.
An increase in the separation of the two peaks indicates an improved signal-tonoise ratio, but an extension in the tails of the PDFs counteracts this effect: the
Bayesian error stays at 10.46%, in agreement with theoretical predictions [9] that
hold for any strictly monotonic map from g8 into Vm . For the "( = 1.4 operational
curve, the PDFs that characterize Vm depend on Pr and are shown in Fig 4C
(Pr = 0.5, for which gp is independently drawn from its prior in each time step) and
4D (Pr = 0.975, which describes phasic input persistance for about 400 ms). The
implementation of Bayesian optimal detection of gp -::f- 0 for "( = 1.4 requires three
separating boundaries; the corresponding Bayesian errors stand at 10.46% for Fig
4C and 4.23% for Fig 4D. A single separating boundary in the gap between the two
stable branches is suboptimal, but is easily implement able by the bistable neuron.
This strategy leads to detection errors of 20.06% for Fig 4C and 4.38% for Fig 4D.
Note that the Bayesian error decreases only when time correlations are included,
and that in this case, detection based on a single separating boundary is very close
to optimal. The results for "( = 1.4 clearly indicate that ambiguities in the bistable
region make it harder to identify temporally uncorrelated instances of gp -::f- 0 on
the basis of a single separating boundary (Fig 4C), while performance improves if
instances with gp -::f- 0 are correlated over time (Fig 4D). Bistability thus provides a
mechanism for improved detection of time correlated input signals.
5
Conclusions
The model presented here incorporates the most relevant effects of dopamine neuromodulation of striatal medium spiny neurons via D1 receptor activation. In the
absence of dopamine the model reproduces the bimodal character of medium spiny
neurons [5]. In the presence of dopamine, the model undergoes a bifurcation and
becomes bistable. This qualitative change in character provides a mechanism to
account for both enhancement and depression of spiny neuron discharge in response
to inputs associated with expectation of reward. There is only limited direct experimental evidence of bistability in the membrane potential of spiny neurons: the
sustained depolarization observed in vitro following brief current injection in the
presence of D1 agonists [4] is a hallmark of bistable responsiveness.
The activity of single striatal spiny neurons recorded in a memory guided saccade
task [1] is strongly modulated by the expectation of reward as reinforcement for correct performance. In these experiments, most units show a more intense response of
longer duration to the presentation of visual stimuli indicative of upcoming reward;
a few units show instead suppressed activity. These observations are consistent
with properties of the model neuron, which is capable of both types of response to
such stimuli. The model identifies the strength of the total excitatory cortical input
as the experimental parameter that selects between these two response types, and
suggests that enhanced responses can have a range of amplitudes but attenuated
responses result in an almost complete suppression of activity, in agreement with
experimental data [1].
Bistability provides a gain mechanism that nonlinearly amplifies both the intensity
and duration of striatal activity. This amplification, exported through thalamocortical pathways, may provide a mechanism for the preferential cortical encoding of
salient information related to reward acquisition. The model indicates that through
the activation of D1 receptors, dopamine can temporarily desensitize spiny neurons
to weak inputs while simultaneously sensitizing spiny neurons to large inputs. A
computational advantage of this mechanism is the potential adaptability of signal
modulation: the brain may be able to utilize the demonstrated plasticity of corti costriatal synapses so that dopamine release preferentially enhances salient signals
related to reward. This selective enhancement of striatal activity would result in a
more informative efferent signal related to achieving reward.
At the systems level, dopamine plays a significant role in the normal operation
of the brain, as evident in the severe cognitive and motor deficits associated with
pathologies ofthe dopamine system (e.g. Parkinson's disease, schizophrenia). Yet at
the cellular level, the effect of dopamine on the physiology of neurons seems modest.
In our model , a small increase in the magnitude of both IRK and L-Ca currents
elicited by D1 receptor activation suffices to switch the character of spiny neurons
from bimodal to truly bistable, which not only modulates the frequency of neural
responses but also introduces a state dependence and a temporal effect. Other
models have suggested that dopamine modulates contrast [9], but the temporal
effect is a novel aspect that plays an important role in information processing.
6
References
[1] Kawagoe R, Takikawa Y , Hikosaka 0 (1998). Nature Neurosci 1:411-416.
[2] Schultz W (1998). J Neurophysiol 80:1-27.
[3] Nicola SM, Surmeier DJ, Malenka RC (2000). Annu Rev Neurosci 23:185-215.
[4] Hernandez-Lopez S, Bargas J , Surmeier DJ , Reyes A, Gallarraga E (1997) . J Neurosci
17:3334-42.
[5] Wilson CJ, Kawaguchi Y (1996). J Neurosci 7:2397-2410 .
[6] Hille B (1992) Ionic Channels of Excitable Membranes. Sinauer Ass., Sunderland MA.
[7] Pacheco-Cano MT, Bargas J , Hernandez-Lopez S (1996) . Exp Brain Res 110:205-21l.
[8] Surmeier DJ, Bargas J , Hemmings HC , Nairn AC, Greengard P (1995). Neuron 14:385397.
[9] Servan-Schreiber D, Printz H, Cohen JD (1990). Science 249:892-895.
| 2142 |@word trial:10 seems:1 hyperpolarized:6 cm2:3 d2:1 grey:1 gradual:1 t_:1 solid:7 harder:1 initial:1 current:37 surprising:1 activation:12 yet:1 chicago:1 hyperpolarizing:5 subsequent:1 informative:1 shape:1 plasticity:1 motor:2 v:1 alone:1 cue:1 stationary:4 indicative:1 short:2 provides:5 characterization:1 node:1 location:1 detecting:1 rc:1 constructed:2 direct:2 differential:1 become:1 qualitative:1 consists:2 sustained:1 lopez:2 pathway:1 manner:2 expected:4 rapid:2 behavior:4 brain:3 goldman:1 increasing:3 becomes:2 motivational:1 suffice:1 panel:3 medium:3 inward:4 cm:9 monkey:1 depolarization:4 suppresses:1 finding:1 transformation:1 differentiation:1 temporal:5 demonstrates:1 whatever:1 unit:9 consequence:2 striatum:4 receptor:13 encoding:1 firing:6 modulation:4 fluctuation:1 jls:6 might:3 black:1 hernandez:2 suggests:1 sara:1 limited:3 range:7 unique:1 spontaneously:1 implement:1 substantia:1 intersect:1 elicit:3 physiology:1 cannot:1 close:1 context:2 influence:1 map:1 demonstrated:1 duration:5 l:2 independently:1 stabilizing:1 classic:1 notion:2 discharge:3 enhanced:8 target:2 play:3 trigger:2 hypothesis:2 origin:1 agreement:2 asymmetric:1 ark:2 observed:3 role:4 region:1 episode:1 oe:1 solla:1 decrease:4 e8:1 removed:2 substantial:1 disease:1 reward:21 dynamic:1 depend:5 purely:1 basis:1 gu:1 neurophysiol:1 rightward:1 easily:1 represented:1 irk:9 spend:1 valued:2 larger:2 ability:2 gi:3 ward:1 gp:16 emergence:2 noisy:1 advantage:1 biophysical:1 net:1 product:2 relevant:1 realization:1 achieve:1 amplification:2 description:3 amplifies:1 potassium:1 enhancement:7 perfect:1 depending:1 illustrate:1 ac:1 eq:10 indicate:1 met:2 direction:1 guided:2 closely:1 correct:2 bistable:9 separatrix:1 noticeably:1 suffices:1 extension:1 pl:1 strictly:1 hold:1 around:3 normal:1 houk:2 exp:1 equilibrium:5 stabilizes:1 driving:4 major:1 vary:1 g8:34 precede:1 modulating:1 schreiber:1 instantaneously:1 clearly:1 always:1 gaussian:1 pacheco:1 inactivation:1 parkinson:1 varying:1 voltage:6 wilson:1 release:4 derived:1 pdfs:6 indicates:3 contrast:2 suppression:2 baseline:3 detect:1 dependent:3 hille:1 explanatory:1 sunderland:1 reproduce:1 quasi:1 selects:2 selective:1 upward:1 retaining:1 bifurcation:2 equal:1 nairn:1 anticipated:1 report:1 stimulus:6 few:1 simultaneously:1 kawagoe:1 conductance:10 detection:4 investigate:1 severe:1 introduces:2 generically:1 truly:1 activated:1 accurate:1 capable:1 preferential:1 shorter:1 intense:1 modest:1 permeability:1 circle:1 re:1 theoretical:1 minimal:1 instance:3 increased:1 servan:1 bistability:9 depolarizes:2 characterize:1 reported:1 dependency:1 gd:4 combined:1 density:2 peak:2 stay:1 physic:1 vm:30 enhance:3 again:1 ambiguity:1 satisfied:1 recorded:1 opposed:2 slowly:1 cognitive:2 leading:1 return:1 li:1 account:4 potential:20 summarized:1 includes:2 hysteresis:1 satisfy:1 mv:6 depends:3 vi:5 multiplicative:2 later:1 start:1 maintains:1 elicited:2 slope:1 depolarizing:1 vivo:1 rectifying:2 ass:2 il:2 compartment:1 square:1 characteristic:3 subthreshold:1 correspond:2 identify:1 ofthe:1 weak:1 biophysically:1 bayesian:5 accurately:1 agonist:3 hli:1 ionic:7 critically:1 trajectory:1 drive:1 straight:2 explain:2 synapsis:1 reach:5 whenever:1 synaptic:13 acquisition:1 frequency:1 involved:1 james:1 associated:4 attributed:1 static:1 efferent:1 gain:2 improves:2 electrophysiological:1 cj:1 schedule:1 amplitude:4 adaptability:1 higher:2 dt:1 follow:1 response:27 improved:2 formulation:1 though:1 strongly:1 biomedical:1 until:3 correlation:2 hand:3 ei:1 nonlinear:1 logistic:1 undergoes:1 gray:1 indicated:1 effect:16 evolution:2 former:1 steady:2 m:3 prominent:2 pdf:3 evident:1 complete:1 vmf:1 reyes:1 cano:1 meaning:1 hallmark:1 novel:1 mt:1 vitro:1 cohen:1 tail:1 counteracts:2 katz:1 refer:1 significant:1 neuromodulatory:3 consistency:1 depressed:1 cancellation:1 pathology:1 dj:3 stable:7 longer:3 gt:11 playa:1 showed:1 leftward:1 rewarded:7 scenario:1 exported:1 responsiveness:1 additional:3 nigra:1 determine:2 monotonically:2 signal:6 ii:2 relates:1 branch:13 smooth:2 exceeds:3 match:1 characterized:1 hikosaka:1 compensate:1 schizophrenia:1 controlled:1 prediction:1 expectation:3 dopamine:40 grounded:1 bimodal:3 ion:9 cell:1 receive:1 whereas:1 addition:1 background:1 decreased:1 printz:1 operate:1 depolarized:8 exhibited:1 induced:2 simulates:1 incorporates:2 dvm:1 unrewarded:6 presence:4 intermediate:2 exceed:1 switch:3 suboptimal:1 attenuated:1 whether:3 depress:1 abruptly:2 resistance:1 action:2 depression:2 ignored:1 modulatory:2 se:1 clear:1 outward:2 amount:1 dark:1 band:1 inhibitory:1 alters:1 dotted:6 stabilized:1 detectability:1 basal:2 key:1 zv:1 terminology:1 threshold:1 inactivate:1 salient:2 drawn:1 achieving:1 verified:1 utilize:1 year:1 ork:4 respond:1 hodgkin:1 almost:1 separation:4 delivery:2 fl:1 bound:2 followed:3 activity:16 strength:5 occur:2 awake:1 encodes:1 aspect:2 span:1 extremely:1 malenka:1 expanded:1 injection:1 department:1 according:1 hyperpolarizes:3 combination:1 membrane:19 smaller:2 describes:2 slightly:2 suppressed:6 spiny:24 character:5 rev:1 pr:8 equation:4 remains:3 opposes:1 neuromodulation:2 mechanism:7 needed:1 phasic:12 drastic:1 reversal:1 operation:1 greengard:1 generic:1 corti:1 existence:1 jd:1 include:1 nicola:1 kawaguchi:1 suddenly:1 leakage:2 upcoming:1 added:1 occurs:1 strategy:1 primary:1 dependence:4 concentration:1 rt:2 enhances:2 exhibit:1 deficit:1 separating:4 majority:1 unstable:2 cellular:1 modeled:6 providing:1 balance:1 ratio:1 preferentially:1 mostly:1 striatal:6 negative:1 rise:1 suppress:1 design:1 reliably:1 calcium:3 implementation:1 desensitize:1 gated:5 upper:6 neuron:35 observation:2 sm:1 t:3 tonic:3 precise:1 overtakes:1 intensity:2 introduced:1 nonlinearly:1 narrow:1 able:2 suggested:1 dynamical:1 below:2 regime:1 solut:1 memory:2 critical:3 event:1 overlap:1 force:4 predicting:1 representing:2 brief:1 temporally:2 disappears:1 identifies:1 bargas:3 excitable:1 prior:3 literature:1 determining:1 relative:3 sinauer:1 par:1 northwestern:2 interesting:1 generation:1 sufficient:3 consistent:3 gruber:1 principle:1 uncorrelated:1 excitatory:3 course:1 accounted:1 repeat:1 thalamocortical:1 side:3 allow:1 curve:15 boundary:4 cortical:4 transition:6 stand:1 sensory:7 author:2 qualitatively:1 reinforcement:2 schultz:1 compacta:1 overcomes:1 reproduces:1 active:1 assumed:2 un:1 channel:3 tonoise:1 nature:1 ca:9 operational:12 contributes:1 improving:1 as:1 hc:1 da:4 intracellular:1 arrow:4 neurosci:4 noise:2 mediate:1 fig:24 slow:1 lie:1 down:14 annu:1 specific:4 physiological:1 cease:1 evidence:1 modulates:4 magnitude:1 downward:1 conditioned:1 gap:2 intersection:2 likely:1 saddle:1 ganglion:2 appearance:2 visual:3 unexpected:1 adjustment:1 temporarily:1 saccade:2 speculated:1 monotonic:1 corresponds:2 determines:1 ma:1 marked:1 goal:1 presentation:1 towards:5 absence:1 experimentally:1 change:4 included:1 except:2 principal:1 total:8 specie:1 secondary:1 experimental:5 aaron:1 sustainable:1 arises:2 modulated:1 incorporate:1 surmeier:3 d1:11 correlated:6 |
1,255 | 2,143 | A Convergent Form of Approximate Policy
Iteration
Theodore J. Perkins
Department of Computer Science
University of Massachusetts Amherst
Amherst, MA 01003
[email protected]
Doina Precup
School of Computer Science
McGill University
Montreal, Quebec, Canada H3A 2A7
[email protected]
Abstract
We study a new, model-free form of approximate policy iteration which
uses Sarsa updates with linear state-action value function approximation
for policy evaluation, and a ?policy improvement operator? to generate
a new policy based on the learned state-action values. We prove that if
the policy improvement operator produces -soft policies and is Lipschitz
continuous in the action values, with a constant that is not too large, then
the approximate policy iteration algorithm converges to a unique solution from any initial policy. To our knowledge, this is the first convergence result for any form of approximate policy iteration under similar
computational-resource assumptions.
1 Introduction
In recent years, methods for reinforcement learning control based on approximating value
functions have come under fire for their poor, or poorly-understood, convergence properties. With tabular storage of state or state-action values, algorithms such as Real-Time
Dynamic Programming, Q-Learning, and Sarsa [2, 13] are known to converge to optimal
values. Far fewer results exist for the case in which value functions are approximated using generalizing function approximators, such as state-aggregators, linear approximators,
or neural networks. Arguably, the best successes of the field were generated in this way
(e.g., [15]), and there are a few positive convergence results, particularly for the case of linear approximators [16, 7, 8]. However, simple examples demonstrate that many standard
reinforcement learning algorithms, such as Q-Learning, Sarsa, and approximate policy iteration, can diverge or cycle without converging when combined with generalizing function
approximators (e.g., [1, 6, 4]).
One classical explanation for this lack of convergence is that, even if one assumes that the
agent?s environment is Markovian, the problem is non-Markovian from the agent?s point
of view?the state features and/or the agent?s approximator architecture may conspire to
make some environment states indistinguishable. We focus on a more recent observation,
which faults the discontinuity of the action selection strategies usually employed by reinforcement learning agents [5, 10]. If an agent uses almost any kind of generalizing function
approximator to estimate state-values or state-action values, the values that are learned depend on the visitation frequencies of different states or state-action pairs. If the agent?s
behavior is discontinuous in its value estimates, as is the case with greedy and -greedy
behavior [14], then slight changes in value estimates may result in radical changes in the
agent?s behavior. This can dramatically change the relative frequencies of different states
or state-action pairs, causing entirely different value estimates to be learned.
One way to avoid this problem is to ensure that small changes in action values result in
small changes in the agent?s behavior?that is, to make the agent?s policy a continuous
function of its values. De Farias and Van Roy [5] showed that a form of approximate
value iteration which relies on linear value function approximations and softmax policy improvement is guaranteed to possess fixed points. For partially-observable Markov decision
processes, Perkins and Pendrith [10] showed that observation-action values that are fixed
points under Q-Learning or Sarsa update rules are guaranteed to exist if the agent uses
any continuous action selection strategy. Both of these papers demonstrate that continuity
of the agent?s action selection strategy leads to the existence of fixed points to which the
algorithms can converge. In neither case, however, was convergence established.
We take this line of reasoning on step further. We study a form of approximate policy iteration in which, at each iteration: (1) Sarsa updating is used to learn weights for a linear
approximation to the action value function of the current policy (policy evaluation), and
then (2) a ?policy improvement operator? determines a new policy based on the learned
action values (policy improvement).1 We show that if the policy improvement operator,
analogous to the action selection strategy of an on-line agent, is -soft and Lipschitz continuous in the action values, with a constant that is not too large, then the sequence of policies
generated is guaranteed to converge. This technical requirement formalizes the intuition
that the agent?s behavior should not change too dramatically when value estimates change.
2 Markov Decision Processes and Value Functions
! "#
$
&%
*)
$"+
"
#
$
"
'
(
*
)
%
" ,
- - .'0/1/0/ 2
2 12 64
354
9;< : 2
9;< : 2
8
7
8
7
3 4 # , 4 2>=?A@ 2 ? CBD 6 4 # E
4 2>=?A@ 2 ?
?
FBD
@ HG ) %
JI
6K4L
KM
6K4
PO ) QSR
"# E
NM
TVUXW Q R
T"UXW
We consider infinite-horizon discounted Markov decision problems [3]. We assume that
the Markov decision process has a finite state set, , and a finite action set, , with sizes
and
. When the process is in state and the agent chooses action ,
the agent receives an immediate reward with expectation , and the process transitions to
next state with probability
. Let be the length
vector of expected immediate
rewards following each state-action pair (
).
A stochastic policy, , assigns a probability distribution over to each
. The prob. If
ability that the agent chooses action when the process is in state is denoted
is deterministic in state , i.e., if
for some and
for all
,
then we write
. For
let , , and denote, respectively, the state
of the process at time , the action chosen by the agent at time , and the reward received
by the agent at time . For policy , the state-value function,
, and state-action value
function (or just action-value function),
, are defined as:
where the expectation is with respect to the stochasticity of the process and the fact that the
agent chooses actions according to , and
is a discount factor. It is well-known
[11] that there exists at least one deterministic, optimal policy
for which
for all , , and .
Policy is called -soft if
for all and . For any
, let
denote the set
of -soft policies. Note that a policy, , can be viewed as an element of
, and
can
be viewed as a compact subset of
. We make the following assumption:
Assumption 1 Under any policy , the Markov decision process behaves as an irreducible, aperiodic Markov chain over the state set .
1
The algorithm can also be viewed as batch-mode Sarsa with linear action-value function approximation.
?
????????????????????????????????????
Inputs: initial policy , and policy improvement operator .
for i=0,1,2,. . . do
Policy evaluation: Sarsa updates under policy , with linear function approximation.
Initialize
arbitrarily.
With environment in state :
.
Choose according to
Observe ,
.
Repeat for
until converges:
.
Choose according to
T
? # ?
1? ?
-
2 % . 0/1/0/ 2
A2 2
2 0> 2 # 2 E
2
@
12 12
!" V #
Observe
.
# 2 E
2
Policy improvement:
.
end for
????????????????????????????????????
Figure 1: The version of approximate policy iteration that we study.
The approximate policy iteration algorithm we propose learns linear approximations to the
action value functions of policies. For this purpose, we assume that each state-action pair
is represented by a length $ feature vector
. (In this paper, all vectors are
columns
unless transposed.) For weights
, the approximate action-value for
%
is
denotes the transpose of
. Letting be the
& , where
-by- $ matrix whose rows correspond to the feature vectors of the state-action pairs, the
entire
approximate action-value function given by weights is represented by the vector
%
. We make the following assumption:
#
$
6!# E
#
$
6
#
$
T #
$
#
$
E
Assumption 2 The columns of are linearly independent.
?
3
4
34
6 K4 3 4
3 4 3 4# 6 4
3 4 # , M 3K4 # ,
3 (4 '
3 Approximate Policy Iteration
The standard, exact policy iteration algorithm [3] starts with an arbitrary policy
and
alternates between two steps: policy evaluation, in which (' is computed, and pol*' can be computed
icy improvement, in which a new policy, ) , is computed.
in various ways, including dynamic programming or solving a system of linear equations. !" is taken to be a greedy, deterministic policy with respect to (' . That is,
,+.-/0#+21
('
+2-3/0+
1
for all . Policy it)
54
(' . It is well-known that the sequence of policies
eration terminates when (')687
'"697
'
generated is monotonically improving in the sense that
for all , and
that the algorithm terminates after a finite number of iterations [3].
# ,
Bertsekas and Tsitsiklis [4] describe several versions of approximate policy iteration in
which the policy evaluation step is not exact. Instead,
is approximated by a weighted
linear combination of state features, with weights determined by Monte Carlo or TD( : )
learning rules. However, they assume that the policy improvement step is the same as
in the standard policy iteration algorithm?the next policy is greedy with respect to the
(approximate) action values of the previous policy. Bertsekas and Tsitsiklis show that if the
approximation error in the evaluation step is low, then such algorithms generate solutions
that are near optimal [4]. However, they also demonstrate by example that the sequence
of policies generated does not converge for some problems, and that poor performance can
result when the approximation error is high.
We study the version of approximate policy iteration shown in Figure 1. Like the versions
studied by Bertsekas and Tsitsiklis, we assume that policy evaluation is not performed
exactly. In particular, we assume that Sarsa updating is used to learn the weights of a linear
approximation to the action-value function. We use action-value functions instead of statevalue functions so that the algorithm can be performed based on interactive experience
with the environment, without knowledge of the state transition probabilities. The weights
learned in the policy evaluation step converge under conditions specified by Tsitsiklis and
Van Roy [17], one of which is Assumption 2.
6 T UXW
6 6 TJUXW V6 6 #TJ6 UXW # 6 6 6
O )
O )
C?
The key difference from previous work is that we assume a generic policy improvement operator, , which maps every
to a stochastic policy. This operator may produce,
for example, greedy policies, -greedy policies, or policies with action selection probabilities based on the softmax
[14]. is Lipschitz continuous with constant if, for
function
all
,
, where denotes the Euclidean
norm. is -soft if, for all
,
is -soft. The fact that we allow for a policy
improvement step that is not strictly greedy enables us to establish the following theorem.
Theorem 1 For any infinite-horizon Markov decision process satisfying Assumption 1, and
for any
, there exists
such that if is -soft and Lipschitz continuous with
constant , then the sequence of policies generated by the approximate policy iteration
, regardless of the
algorithm in Figure 1 converges to a unique limiting policy
choice of .
QR
In other words, if the behavior of the agent does not change too greatly in response to
changes in its action value estimates, then convergence is guaranteed. The remainder of
the paper is dedicated to proving this theorem. First, however, we briefly consider what the
theorem means and what some of its limitations are. The strength of the theorem is that it
states a simple condition under which a form of model-free reinforcement learning control
based on approximating value functions converges for a general class of problems. The
theorem does not specify a particular constant, , which ensures convergence; it merely
states that such a constant exists. The values of (and hence, range of policy improvement
operators) which ensure convergence depend on properties of the decision process, such
as its transition probabilities and rewards, which we assume to be unknown. The theorem
also offers no guarantee on the quality of the policy to which the algorithm converges.
Intuitively, if the policy improvement operator is Lipschitz continuous with a small constant
, then the agent is limited in the extent to which it can optimize its behavior. For example,
even if an agent correctly learns that the value of action is much higher than the value
of action , limits the frequency with which the agent can choose in favor of , and
this may limit performance. The practical importance of these considerations remains to
be seen, and is discussed further in the conclusions section.
$
4 Proof of Theorem 1
4.1 Probabilities Related to State-Action Pairs
K4
2
# ,
( 2
4 "# E
$
, K4 K4
4
# E
# ,#E
$
4
> 1
(
# #E
$ E # #E
$ #
$ 0+
1 1#E
$(
# #E
$( :
Because the approximate policy iteration algorithm in Figure 1 approximates action-values,
our analysis relies extensively on certain probabilities that are associated with state-action
pairs. First, we define
to be the
-bymatrix whose entries correspond to the
probabilities that one state-action pair follows another when the agent behaves according
row and
column of is
to . That is, the element on the
.
can be viewed as the stochastic transition matrix of a Markov chain over state-action
pairs.
.
Lemma 1 There exists
such that for all , 7
Proof: Let and be fixed, and let
and
. Then 7
4
. It is readily shown that for any two -by- matrices
K4 7 K 4
4 O ) # E
2
4
4
"
#
E
4
"# AE4
O )
F4
4
O )
E Q R 427 4
QPR :C4
4
:C4
:
0+
1 4
: 4 :C4 %
QPR Q R
:
7
:
Q R UX W 427 4 427 4 UX W 7 K427 K4 : UX W K4
7 K 4
!!4 7 ! ! 4 4 4 7
4 A4
and
whose elements different in absolute value by at most ,
.
. Hence,
Under Assumption 1, fixing a policy, , induces an irreducible, aperiodic Markov chain
over . Let
denote the stationary probability of state . We define
to be
the length
vector whose
element is
. Note that the elements of
sum to one. If
for all and , then all elements of
are positive and it
is easily verified that
is the unique stationary distribution of the irreducible, aperiodic
Markov chain over state-action pairs with transition matrix
.
Lemma 2 For any
.
, there exists
such that for all
,
Proof: For any
, let
be the largest eigenvalue of
with modulus strictly less
than 1.
is well-defined since the transition matrix of any irreducible, aperiodic Markov
chain has precisely one eigenvalue equal to one [11]. Since the eigenvalues of a matrix
are continuous in the elements of the matrix [9], and since
is compact, there exists
for some
. Seneta [12], showed that for any
and
and stationary
two irreducible aperiodic Markov chains with transition matrices
distributions
and , on a state set with elements,
,
where
is the largest eigenvalue of
with modulus strictly less than one. Let
.
.
Lastly, we define
,
to be the matrix whose diagonal is
.
. It is easy to show that for any
4.2 The Weights Learned in the Policy Evaluation Step
4
! 4
F2
! 4 " 4 ! 4
@ :
F4
#!!4$" K4 % 4 #!!4$
N
4
@
N4
4
N4' % 4
% 4 7 % 4
(&
427 4 )' & ('
% 24 7 % 4 # !!24 7 !!4 !!427 ! 4
)
427 4 G ! 24 7 4 $ 7" @ 24 7 4 7 ! 4 4
$ " @ 4 4
+ *
!! 4 7 $ " !
4@ ! 4 ! 7 4 7$" @ !
4
4
! 24 7 ! 4 @ !
4 7 24 7 @ 4 4 ! 4 4
@
@
" E
O )
Consider the approximate policy evaluation step of the algorithm in Figure 1. Suppose
that the agent follows policy and uses Sarsa updates to learn weights , and suppose
that
for all and . Then
is the stochastic transition matrix of an irreducible, aperiodic Markov chain over state-action pairs, and
has the unique stationary
distribution of that chain on its diagonal. Under standard conditions on the learning rate
parameters for the updates, , Tsitsiklis and Van Roy [17] show that the weights converge
to the unique solution to the equation:
(1)
(Note that we have translated their result for TD( ) updating of approximate state-values
to Sarsa, or TD(0), updating of approximate state-action values.) In essence, this equation
says that the ?expected update? to the weights under the stationary distribution, , is zero.
Let
and
. Tsitsiklis and Van Roy [17] show that
is invertible, hence we can write
for the unique weights which satisfy
Equation 1.
Lemma 3 There exist
and
such that for all
.
Proof: For the first claim,
. For the second claim,
,
and
4 7 ! 4
@ ! 4 7 4 7 4
@ $! 4 7 ! 4
4
% 42 7 ! "4 @ ! 42 7 427 4 @ ! 427 ! 4 4
! 4 %
4 %
@ @
QPR O )
!4
QR
4
4
S
Q
R
TUW
0 =
O)
O)
QPR N4' M
DQ R N4' M 4 7
QP R 427 4' M
S
Q
R
N4 E O ) PO )
S4
E Q R !4 7 !4
E Q R
P24 7 !24 7 % 24 7 N4 !4 % 4
24 7
4 7 4 4
24 7
4
4 7 24 7 4 4 4 4 %% 24 7 %% 4
4 7 4 7 4
# 4 47 7 4 7 4
4 4
% 4 4 7 7 % 4
4 4 7
4
4
24 7 24 7 4 % % 24 7 % % 4 24 7 4 & 4
24 7 4 &
'
24 7 4 (& )'
!
!
where the last line follows from Lemmas 1 and 2 and the facts
for any
.
, there exists such that
Lemma 4 For any
and
for all
.
Proof: By Lemmas 1 and 2, and by the continuity of matrix inverses [11],
is a continuous function of . Thus,
is a continuous function of . Because
is a compact
subset of
, and because continuous functions map compact sets to compact sets, the
existence of the bound, , follows.
For any
-bymatrix , let
. That is,
measures
how
small a vector of length one can become under left-multiplication by matrix .
, there exists
Lemma 5 For any
,
such that for all
.
Proof: Lemma 7, in the Appendix, shows that is a continuous mapping and that is
,
. Since
positive for any non-singular matrix. For any
is continuous, and
compact, the infimum is attained by some
. Thus
, where the last inequality follows because is non-singular.
Lemma 6 For any
.
Proof: Let
Thus:
, there exists
such that for all
,
be arbitrary. From Equation 1,
and
.
(2)
The left hand side of Equation 2 follows from Lemmas 5 and 7; the right hand side follows
from Lemmas 3 and 4.
4.3 Contraction Argument
O )
V 6 4 ! 4 ! 4 6 4
Proof of Theorem 1: For a given infinite-horizon discounted Markov decision problem,
let
and be fixed. Suppose that is Lipschitz continuous with constant , where
is yet to be determined. Let
be arbitrary. The policies that result from
and% after one% iteration of the approximate policy% iteration % algorithm of% Figure% 1 are
7
respectively.
7
Observe that:
7 2and
from Lemma 6. If
, where the% last step follows
%
7 , then for some
7
! . Each
we have
!
iteration of the approximate policy iteration is a contraction. By the
% Contraction Mapping
Theorem [3], there is a unique fixed point of the mapping #"
, and the sequence of
policies generated according to that mapping from any initial policy converges to the fixed
point.
6K4
E Q R
4
4
4
4
V
6
6
6
6
G ) %
6 4 V 6 4
V 6 4$
K6 % 4
Note that since the sequence of policies, , converges, and since
is a continuous function
of
,
the
sequence
of
approximate
action-value
functions
computed
by the algorithm,
%
' , also converges.
5 Conclusions and Future Work
We described a model-free, approximate version of policy iteration for infinite-horizon discounted Markov decision problems. In this algorithm, the policy evaluation step of classical
policy iteration is replaced by learning a linear approximation to the action-value function
using on-line Sarsa updating. The policy improvement step is given by an arbitrary policy
improvement operator, which maps any possible action-value function to a new policy. The
main contribution of the paper is to show that if the policy improvement operator is -soft
and Lipschitz continuous in the action-values, with a constant that is not too large, then
the approximate policy iteration algorithm is guaranteed to converge to a unique, limiting
policy from any initial policy. We are hopeful that similar ideas can be used to establish the
convergence of other reinforcement learning algorithms, such as on-line Sarsa or Sarsa( : )
control with linear function approximation.
The magnitude of the constant that ensures convergence depends on the model of the
environment and on properties of the feature representation. If the model is not known, then
choosing a policy improvement operator that guarantees convergence is not immediate.
To be safe, an operator for which is small should be chosen. However, one generally
prefers to be large, so that the agent can exploit its knowledge by choosing actions with
higher estimated action-values as frequently as possible. One approach to determining
a proper value of would be to make an initial guess and begin the approximate policy
iteration procedure. If the contraction property fails on any iteration, one should choose a
new policy improvement operator that is Lipschitz continuous with a smaller constant. A
potential advantage of this approach is that one can begin with a high choice of , which
allows exploitation of action value differences, and switch to lower values of only as
necessary. It is possible that convergence could be obtained with much higher values of
than are suggested by the bound in the proof of Theorem 1.
Discontinuous improvement operators/action selection strategies can lead to nonconvergent behavior for many reinforcement learning algorithms, including Q-Learning,
Sarsa, and forms of approximate policy iteration and approximate value iteration. For
some of these algorithms, (non-unique) fixed points have been shown to exist when the
action selection strategy/improvement operator is continuous [5, 10]. Whether or not convergence also follows remains to be seen. For the algorithm studied in this paper, we have
constructed an example demonstrating non-convergence with improvement operators that
are Lipschitz continuous but with too large of a constant. In this case, it appears that the
Lipschitz continuity assumption we use cannot be weakened. One direction for future work
is determining minimal restrictions on action selection (if any) that ensure the convergence
of other reinforcement learning algorithms.
Ensuring convergence answers one standing objection to reinforcement learning control
methods based on approximating value functions. However, an important open issue for
our approach, and for other approaches advocating continuous action selection [5, 10], is
to characterize the solutions that they produce. We know of no theoretical guarantees on
the quality of solutions found, and there is little experimental work comparing algorithms
that use continuous action selection with those that do not.
Acknowledgments
Theodore Perkins was supported in part by National Science Foundation grants ECS0070102 and ECS-9980062. Doina Precup was supported in part by grants from NSERC
and FQNRT.
References
[1] L. C. Baird. Residual algorithms: Reinforcement learning with function approximation. In Proceedings of the Twelfth International Conference on Machine Learning, pages 30?37. Morgan
Kaufmann, 1995.
[2] A. G. Barto, S. J. Bradtke, and S. P. Singh. Learning to act using real-time dynamic programming. Artificial Intelligence, 72(1):81?138, 1995.
[3] D. P. Bertsekas. Dynamic Programming and Optimal Control, Volumes 1 and 2. Athena
Scientific, 2001.
[4] D. P. Bertsekas and J. N. Tsitsiklis. Neuro-Dynamic Programming. Athena Scientific, 1996.
[5] D. P. De Farias and B. Van Roy. On the existence of fixed points for approximate value iteration
and temporal-difference learning. Journal of Opt. Theory and Applications, 105(3), 2000.
[6] G. Gordon. Chattering in Sarsa( ).
www.cs.cmu.edu/ ggordon, 1996.
CMU Learning Lab Internal Report. Available at
[7] G. Gordon. Approximate Solutions to Markov Decision Processes. PhD thesis, Carnegie Mellon
University, 1999.
[8] G. J. Gordon. Reinforcement learning with function approximation converges to a region.
Advances in Neural Information Processing Systems 13, pages 1040?1046. MIT Press, 2001.
[9] C. D. Meyer. Matrix Analysis and Applied Linear Algebra. SIAM, 2000.
[10] T. J. Perkins and M. D. Pendrith. On the existence of fixed points for Q-learning and Sarsa in
partially observable domains. In Proceedings of the Nineteenth International Conference on
Machine Learning, 2002.
[11] M. L. Puterman. Markov Decision Processes: Disrete Stochastic Dynamic Programming. John
Wiley & Sons, Inc, New York, 1994.
[12] E. Seneta. Sensitivity analysis, ergodicity coefficients, and rank-one updates for finite markov
chains. In W. J. Stewart, editor, Numerical Solutions of Markov Chains. Dekker, NY, 1991.
[13] S. Singh, T. Jaakkola, M. L. Littman, and C. Szepesvari. Convergence results for single-step
on-policy reinforcement-learning algorithms. Machine Learning, 38(3):287?308, 2000.
[14] R. S. Sutton and A. G. Barto. Reinforcement Learning: An Introduction. MIT Press/Bradford
Books, Cambridge, Massachusetts, 1998.
[15] G. J. Tesauro. TD-Gammon, a self-teaching backgammon program, achieves master-level play.
Neural Computation, 6(2):215?219, 1994.
[16] J. N. Tsitsiklis and B. Van Roy. Optimal stopping of markov processes: Hilbert space theory,
approximation algorithms, and an application to pricing high-dimensional financial derivatives.
IEEE Transactions on Automatic Control, 44(10):1840?1851, 1999.
[17] J. N. Tsitsiklis and B. Van Roy. An analysis of temporal-difference learning with function
approximation. IEEE Transactions on Automatic Control, 42(5):674?690, 1997.
Appendix
MO ))
T
Lemma 7 For -by- matrix
1.
2.
3.
4.
0
M
, let
for all ,
iff is non-singular,
for any
,
is continuous.
=
. Then:
,
% .' '1/0/1/
I I = I0 I
:
I :
)
I
"0
: 0 = "0
: I )0
: I
all
Ilet 1 I +2-3/ )00
= : . Then
I )0 I : I I "0 I0 3. Now,
for
:
"0 : I I1 M )0 : I I1 M
"0 3 : I I I I1 . Thus, )0 : I0 .
Proof: The first three points readily follow from elementary arguments.
We focus on
the last point. We want to show that given a sequence
of
matrices
,
"
that converge
to
some
,
then
.
Note
that
"
means
that
"0
+.-/ 0
. Let
. Then )0
| 2143 |@word exploitation:1 briefly:1 version:5 norm:1 twelfth:1 open:1 dekker:1 km:1 contraction:4 initial:5 uma:1 current:1 comparing:1 yet:1 readily:2 john:1 numerical:1 enables:1 update:7 stationary:5 greedy:7 fewer:1 guess:1 intelligence:1 constructed:1 become:1 prove:1 dprecup:1 expected:2 behavior:8 frequently:1 discounted:3 td:4 little:1 begin:2 what:2 kind:1 formalizes:1 guarantee:3 temporal:2 every:1 act:1 interactive:1 exactly:1 control:7 qpr:4 grant:2 arguably:1 bertsekas:5 positive:3 understood:1 limit:2 sutton:1 studied:2 weakened:1 theodore:2 limited:1 range:1 unique:9 practical:1 acknowledgment:1 procedure:1 word:1 gammon:1 ilet:1 cannot:1 selection:10 operator:17 storage:1 optimize:1 restriction:1 deterministic:3 map:3 www:1 regardless:1 assigns:1 rule:2 financial:1 proving:1 analogous:1 limiting:2 mcgill:2 play:1 suppose:3 exact:2 programming:6 us:4 element:8 roy:7 approximated:2 particularly:1 updating:5 satisfying:1 region:1 ensures:2 cycle:1 intuition:1 environment:5 pol:1 reward:4 littman:1 dynamic:6 depend:2 solving:1 singh:2 algebra:1 f2:1 farias:2 translated:1 po:2 easily:1 represented:2 various:1 describe:1 monte:1 artificial:1 choosing:2 whose:5 nineteenth:1 say:1 ability:1 favor:1 sequence:8 eigenvalue:4 advantage:1 propose:1 remainder:1 causing:1 iff:1 poorly:1 qr:2 convergence:17 requirement:1 produce:3 converges:9 radical:1 montreal:1 fixing:1 school:1 received:1 c:3 come:1 direction:1 safe:1 aperiodic:6 discontinuous:2 f4:2 stochastic:5 opt:1 elementary:1 sarsa:16 strictly:3 mapping:4 mo:1 claim:2 achieves:1 a2:1 purpose:1 largest:2 weighted:1 mit:2 avoid:1 barto:2 jaakkola:1 focus:2 improvement:22 rank:1 backgammon:1 greatly:1 sense:1 stopping:1 i0:3 entire:1 i1:3 issue:1 denoted:1 k6:1 softmax:2 initialize:1 field:1 equal:1 tabular:1 future:2 report:1 gordon:3 few:1 irreducible:6 national:1 replaced:1 fire:1 evaluation:11 hg:1 chain:10 necessary:1 experience:1 unless:1 euclidean:1 theoretical:1 minimal:1 column:3 soft:8 markovian:2 stewart:1 subset:2 entry:1 too:6 characterize:1 answer:1 combined:1 chooses:3 international:2 amherst:2 siam:1 sensitivity:1 standing:1 diverge:1 invertible:1 precup:2 thesis:1 nm:1 choose:4 book:1 derivative:1 potential:1 de:2 coefficient:1 baird:1 inc:1 satisfy:1 doina:2 depends:1 performed:2 view:1 lab:1 start:1 contribution:1 kaufmann:1 correspond:2 carlo:1 aggregator:1 pendrith:2 frequency:3 proof:10 associated:1 transposed:1 massachusetts:2 knowledge:3 hilbert:1 appears:1 higher:3 attained:1 follow:1 response:1 specify:1 just:1 ergodicity:1 lastly:1 until:1 hand:2 receives:1 a7:1 lack:1 continuity:3 mode:1 infimum:1 quality:2 scientific:2 pricing:1 modulus:2 hence:3 puterman:1 indistinguishable:1 self:1 essence:1 demonstrate:3 dedicated:1 bradtke:1 reasoning:1 consideration:1 behaves:2 ji:1 qp:1 volume:1 discussed:1 slight:1 approximates:1 mellon:1 cambridge:1 automatic:2 teaching:1 stochasticity:1 recent:2 showed:3 tesauro:1 certain:1 inequality:1 arbitrarily:1 success:1 fault:1 approximators:4 seen:2 morgan:1 employed:1 converge:8 monotonically:1 technical:1 h3a:1 offer:1 ensuring:1 converging:1 neuro:1 expectation:2 cmu:2 iteration:30 want:1 objection:1 singular:3 tuw:1 posse:1 quebec:1 near:1 easy:1 switch:1 architecture:1 idea:1 whether:1 york:1 action:57 prefers:1 dramatically:2 generally:1 s4:1 discount:1 extensively:1 induces:1 generate:2 exist:4 estimated:1 correctly:1 write:2 carnegie:1 visitation:1 key:1 demonstrating:1 k4:12 neither:1 conspire:1 verified:1 merely:1 year:1 sum:1 prob:1 inverse:1 master:1 almost:1 decision:11 appendix:2 entirely:1 bound:2 guaranteed:5 convergent:1 strength:1 precisely:1 perkins:5 argument:2 department:1 according:5 alternate:1 combination:1 poor:2 terminates:2 smaller:1 son:1 n4:6 intuitively:1 taken:1 resource:1 equation:6 remains:2 know:1 letting:1 end:1 available:1 observe:3 generic:1 batch:1 existence:4 assumes:1 denotes:2 ensure:3 a4:1 qsr:1 exploit:1 cbd:1 establish:2 approximating:3 classical:2 strategy:6 fbd:1 diagonal:2 athena:2 extent:1 length:4 seneta:2 proper:1 policy:92 unknown:1 observation:2 markov:20 finite:4 immediate:3 arbitrary:4 canada:1 pair:11 specified:1 c4:3 advocating:1 learned:6 icy:1 established:1 discontinuity:1 suggested:1 usually:1 program:1 including:2 explanation:1 eration:1 residual:1 multiplication:1 determining:2 relative:1 limitation:1 approximator:2 foundation:1 agent:26 dq:1 editor:1 row:2 repeat:1 last:4 free:3 transpose:1 supported:2 tsitsiklis:9 side:2 allow:1 absolute:1 hopeful:1 van:7 transition:8 reinforcement:12 far:1 ec:1 transaction:2 approximate:30 observable:2 compact:6 continuous:22 ggordon:1 learn:3 szepesvari:1 ca:1 improving:1 domain:1 main:1 linearly:1 ny:1 wiley:1 uxw:2 fails:1 meyer:1 learns:2 theorem:11 e4:1 nonconvergent:1 exists:9 importance:1 phd:1 magnitude:1 horizon:4 generalizing:3 nserc:1 chattering:1 ux:4 partially:2 determines:1 relies:2 ma:1 viewed:4 lipschitz:10 change:9 infinite:4 determined:2 lemma:13 called:1 bradford:1 experimental:1 internal:1 |
1,256 | 2,144 | Generalized 2 Linear 2 Models
Geoffrey J. Gordon
[email protected]
Abstract
We introduce the Generalized 2 Linear 2 Model, a statistical estimator which combines features of nonlinear regression and factor analysis. A (GL)2M approximately decomposes a rectangular matrix
X into a simpler representation j(g(A)h(B)). Here A and Bare
low-rank matrices, while j, g, and h are link functions. (GL)2Ms
include many useful models as special cases, including principal
components analysis, exponential-family peA, the infomax formulation of independent components analysis, linear regression, and
generalized linear models. They also include new and interesting
special cases, one of which we describe below. We also present an
iterative procedure which optimizes the parameters of a (GL)2M.
This procedure reduces to well-known algorithms for some of the
special cases listed above; for other special cases, it is new.
1
Introduction
Let the m x n matrix X contain an independent sample from some unknown distribution. Each column of X represents a training example, and each row represents a
measured feature of the examples. It is often reasonable to assume that some of the
features are redundant , that is, that there exists a reduced set of I features which
contains all or most of the information in X.
If the reduced features are linear functions of the original features and the distributions of the elements of X are Gaussian, redundancy means we can write X as
the product of two smaller matrices U and V with small sum of squared errors.
This factorization is essentially a singular value decomposition: U must span the
first I dimensions of the left principal subspace of X, while V T must span the first
I dimensions of the right principal subspace. (Since the above requirements do not
uniquely determine U and V, the SVD traditionally imposes additional restrictions
which we will ignore in this paper.)
The SVD has a long list of successes in machine learning, including information
retrieval applications such as latent semantic analysis [1] and link analysis [2]; pattern recognition applications such as "eigenfaces" [3]; structure from motion algorithms [4]; and data compression tools [5]. Unfortunately, the SVD makes two
assumptions which can limit its accuracy as a learning tool.
The first assumption is the use of the sum of squared errors between X and UV as
a loss function. Squared error loss means that predicting 1000 when the answer is
1010 is as bad as saying -7 when the answer is 3. The second assumption is that
the reduced features are linear functions of the original features. Instead, X might
be a nonlinear function of UV, and U and V might be nonlinear functions of some
other matrices A and B. To address these shortcomings, we propose the model
x = f(g(A)h(B))
(1)
for the expected value of X. We also propose allowing non-quadratic loss functions
for the error (X - X) and the parameter matrices A and B. The fixed functions
are called link functions. By analogy to generalized linear models [6], we call equation (1) a Generalized 2 Linear 2 Model: generalized2 because it uses link functions
for the parameters A and B as well as the prediction X, and linear2 because like
the SVD it is bilinear.
As long as we choose link and loss functions that match each other (see below for the
definition of matching link and loss), there will exist efficient algorithms for finding
A and B given X, f, g, and h. Because (1) is a generalization of the SVD, (GL)2Ms
are drop-in replacements for SVDs in all of the applications mentioned above, with
better reconstruction performance when the SVD's error model is inaccurate. In
addition, they open up new applications (see section 7 for one) where an SVD would
have been unable to provide a sufficiently accurate reconstruction.
2
Matching link and loss functions
Whenever we try to optimize the predictions of a nonlinear model, we need to worry
about getting stuck in local minima. One example of this problem is when we try
to fit a single sigmoid unit with parameters (J E lRd to training inputs Xi E lRd and
target outputs Yi E lR under squared error loss:
Yi = 10git(Zi)
Zi
= Xi
. (J
Even for small training sets, the number of local minima of L can grow exponentially
with the dimension d [7]. On the other hand, if we optimize the same predictions
Yi under the logarithmic loss function ~i[Yi log Yi + (1 - Vi) 10g(1 - Yi)] instead of
squared error, our optimization problem is convex. Because the logistic link works
with the log loss to produce a convex optimization problem, we say they match each
other [7]. Matching link-loss pairs are important because minimizing a convex loss
function is usually far easier than minimizing a non convex one.
We can use any convex function F(z) to generate a matching pair of link and loss
functions. The loss function which corresponds to F is
(2)
where F*(y) is defined so that minz DF(Z I y) = O. (F* is the convex dual of F [8],
and D F is the generalized Bregman divergence from Z to Y [9].)
Expression (2) is nonnegative, and it is globally convex in all of the ZiS (and therefore
also in (J since each Zi is a linear function of (J). If we write f for the gradient of F,
the derivative of (2) with respect to Zi is f(Zi) - Vi. So, (2) will be zero if and only
if Yi = f(Zi) for all i; in other words, using the loss (2) implies that Yi = f(z;) is
our best prediction of Vi, and f is therefore our matching link function.
We will need two facts about convex duals below. The first is that F* is always
convex, and the second is that the gradient of F* is equal to f - l. (Also, convex
duality is defined even when F, G, and H aren't differentiable. If they are not,
replace derivatives by subgradients below.)
3
Loss functions for (G L )2Ms
In (GL)2Ms, matching loss functions will be particularly important because we need
to deal with three separate nonlinear link functions. We will usually not be able
to avoid local minima entirely; instead, the overall loss function will be convex in
some groups of parameters if we hold the remaining parameters fixed.
We will specify a (GL)2M by picking three link functions and their matching loss
functions. We can then combine these individual loss functions into an overall loss
function as described in section 4; fitting a (GL)2M will then reduce to minimizing
the overall loss function with respect to our parameters. Each choice of links results
in a different (G L)2M and therefore potentially a different decomposition of X.
The choice of link functions is where we should inject our domain knowledge about
what sort of noise there is in X and what parameter matrices A and B are a priori
most likely. Useful link functions include f (x) = x (corresponding to squared error
and Gaussian noise), f (x) = log x (unnormalized KL-di vergence and Poisson noise),
and f(x) = (1 + e- x) - l (log-loss and Bernoulli noise).
The loss functions themselves are only necessary for the analysis; all of our algorithms need only the link functions and (in some cases) their derivatives. So, we
can pick the loss functions and differentiate to get the matching link functions; or,
we can pick the link functions directly and not worry about the corresponding loss
functions. In order for our analysis to apply, the link functions must be derivatives
of some (possibly unknown) convex functions.
Our loss functions are D F , DG, and DH where
G : lRmxl H lR
are convex functions. We will abuse notation and call F, G, and H loss functions as
well: F is the prediction loss, and its derivative f is the prediction link; it provides
our model of the noise in X. G and H are the parameter losses, and their derivatives
g and h are the parameter links; they tell us which values of A and B are a priori
most likely. By convention, since F takes an m x n matrix argument , we will say
that the input and output to f are also m x n matrices (similarly for g and h).
4
The model and its fixed point equations
We will define a (GL)2M by specifying an overall loss function which relates the
parameter matrices A and B to the data matrix X. If we write U = g(A) and
V = h(B), the (GL)2M loss function is
L(U, V) = F(UV) - X
Here A
0
0
UV
+ G*(U) + H*(V)
(3)
B is the "matrix dot product," often written tr(AT B).
Expression (3) is a sum of three Bregman divergences, ignoring terms which don't
depend on U and V: it is DF(UV I X)+DG(O I U) +DH(O I V). The F-divergence
tends to pull UV towards X, while the G- and H-divergences favor small U and V.
To further justify (3), we can examine what happens when we compute its derivatives with respect to U and V and set them to O. The result is a set of fixed-point
equations that the optimal parameter settings must satisfy:
UT(X - f(UV))
B
(4)
(X - f(UV))VT
A
(5)
To understand these equations, we can consider two special cases. First, if we let
G* go to zero (so there is no pressure to keep U and V small) , (4) becomes
UT(X - f(UV)) = 0
(6)
which tells us that each column of the error matrix must be orthogonal to each
column of U. Second, if we set the prediction link to be f(UV) = UV, (6) becomes
UTUV = UTX
which tells us that for a given U, we must choose V so that UV reconstructs X
with the smallest possible sum of squared errors.
5
Algorithms for fitting (GL)2Ms
We could solve equations (4- 5) with any of several different algorithms. For example, we could use gradient descent on either U, V or A, B. Or, we could use the
generalized gradient descent [9] update rule (with learning rate a):
A +-", (X - f(UV))V T
B +-", UT(X - f(UV))
The advantage of these algorithms is that they are simple to implement and don't
require additional assumptions on F , G , and H. They can even work when F, G,
and Hare nondifferentiable by using subgradients.
In this paper, though, we will focus on a different algorithm. Our algorithm is based
on Newton's method , and it reduces to well-known algorithms for several special
cases of (GL)2Ms. Of course, since the end goal is solving (4-5), this algorithm will
not always be the method of choice; instead, any given implementation of a (GL)2M
should use the simplest algorithm that works.
For our Newton algorithm we will need to place some restrictions on the prediction
and parameter loss functions. (These restrictions are only necessary for the Newton
algorithm; more general loss functions still give valid (GL)2Ms, but require different
algorithms.) First, we will require (4-5) to be differentiable. Second, we will restrict
F(Z) = LFij (Zij)
H(B) = L Hj(B. j )
ij
j
These definitions fix most of the second derivatives of L(U, V) to be zero, simplifying
and speeding up computation. Write f ij , gi, and h j for the respective derivatives.
With these restrictions, we can linearize (4) and (5) around our current guess at
the parameters, then solve the resulting equations to find updated parameters. To
simplify notation, we can think of (4) as j separate equations, one for each column
of V. Linearizing with respect to Vj gives:
(U T DjU + Hj)(Vr w
-
Vj) = UT(X.j - f.j(UV j )) - B. j
where the l x l matrix H j is the Hessian of Hi at V j ' or equivalently the inverse of
the Hessian of Hj at B.j; and the m x m diagonal matrix Dj contains the second
derivatives of F with respect to the jth column of its argument. That is,
Now , collecting terms involving Vjew yields:
We can recognize (7) as a weighted least squares problem with weights
precision H j , prior mean Vj + H j1 B-j , and target outputs
UV j
+ Dj1(x.j
VJ5j,
prior
- f(UV j ))
Similarly, we can linearize with respect to rows of U to find the equation
UreW(VDiVT + G i ) = ((Xi. - j;.(Ui.V))Di1 + Ui.V)DiV T + Ui. G i - Ai. (8)
where G i is the Hessian of Gi and Di contains the second derivatives of F with
respect to the ith row of its argument.
We can solve one copy of (7) simultaneously for each column of V, then replace V
by vnew. Next we can solve one copy of (8) simultaneously for each row of U, then
replace U by unew. Alternating between these two updates will tend to reduce (3).1
6
Related models
There are many important special cases of (GL)2Ms. We derive two in this section;
others include principal components analysis, "sensible" PCA, linear regression,
generalized linear models, and the weighted majority algorithm. (Our Newton algorithm turns into power iteration for PCA and iteratively-reweighted least squares
for GLMs.) (GL)2Ms are related to generalized bilinear models; the latter include
some of the above special cases, but not ICA, weighted majority, or the example of
section 7. There are natural generalizations of (GL)2Ms to multilinear interactions.
Finally, some models such as non-negative matrix factorization [10] and generalized low-rank approximation [11] are cousins of (GL)2Ms: they use a loss function
which is convex in either factor with the other fixed but which is not a Bregman
divergence.
6.1
Independent components analysis
In ICA, we assume that there is a hidden matrix V (the same size as X) of independent random variables, and that X was generated from V by applying a square
matrix U. We seek to recover the mixing matrix U and the sources V; in other
words , we want to decompose X = UV so that the elements of V are as nearly
independent as possible.
The info max algorithm for ICA assumes that the elements of V follow distributions
with heavy tails (i.e. , high kurtosis). This assumption helps us find independent
components because the sum of two heavy-tailed random variables tends to have
lighter tails, so we can search for U by trying to make the elements of V follow a
heavy-tailed distribution.
In our notation, the fixed point of the info max algorithm for ICA is
_ U T = tanh(V)XT
(9)
(see, e.g., equation (11) or (13) of [12]). To reproduce (9) , we will let the prediction
link f be the identity, and we will let the duals of the parameter loss functions be
G*(U)
-dogdet U
H*(V)
E
L log cosh
Vij
ij
iTo guarantee convergence, we can check that (3) decreases and reduce our step size if
we encounter problems. (Since U T D j U, H j , V Di V T, and G i are all positive definite, the
Newton update directions are descent directions; so, there always exists a small enough
step size.) We have not found this check necessary in practice.
where
f
is a small positive real number. Then equations (4) and (5) become
UT(X - UV)
(X - UV)VT
ttanh(V)
-fU - T
(10)
(11)
since the derivative of log cosh v is tanh v and the derivative of log det U is U - T .
Right-multiplying (10) by (UV)T and substituting in (11) yields
_u T
Now since UV -+ X as
6.2
f
= tanh(V)(UV)T
(12)
-+ 0, (12) is equivalent to (9) in the limit of vanishing
f.
Exponential family peA
To duplicate exponential family PCA [13], we can set the prediction link f arbitrarily and let the parameter links 9 and h be large multiples of the identity. Our
Newton algorithm is applicable under the assumptions of [13], and (7) becomes
(13)
Equation (13) along with the corresponding modification of (8) should provide a
much faster algorithm than the one proposed in [13], which updates only part of U
or V at a time and keeps updating the same part until convergence before moving
on to the next one.
7
Example: robot belief states
Figure 1 shows a map of a corridor in the CMU CS building. A robot navigating
in this corridor can sense both side walls and compute an accurate estimate of its
lateral position. Unless it is near an observable feature such the lab door near the
middle of the corridor, however, it can't accurately resolve its position along the
corridor and it can't tell whether it is pointing left or right.
In order to plan to achieve a goal in this environment, the robot must maintain
a belief state (a probability distribution representing its best information about
the unobserved state variables). The map shows the robot's starting belief state:
it is at one end of the corridor facing in, but it doesn't know which end. We
collected a training set of 400 belief states by driving the robot along the corridor and
feeding its sensor readings to a belief tracker [14]. To simulate a larger environment
with greater uncertainty, we artificially reduced sensor range and increased error.
Figure 1 shows two of the collected beliefs.
Planning is difficult because belief states are high-dimensional: even in this simple
world there are 550 states (275 positions at lOcm intervals along the corridor x 2
orientations), so a belief is a vector in ]R550. Fortunately, the robot never encounters
most belief states. This regularity can make planning tractable: if we can identify
a few features which extract the important information from belief states, we can
plan in low-dimensional feature space instead of high-dimensional belief space.
We factored the matrix of belief states using feature space ranks l = 3,4, 5. For the
prediction link f(Z) we used exp(Z) (componentwise); this link ensures that the
predicted beliefs are positive, and treats errors in small probabilities as proportionally more important than errors in large ones. (The matching loss for f is a Poisson
log-likelihood or unnormalized KL-divergence.) For the parameter link h we used
10 12 I, corresponding to H* = lO - 12 11V11 2 /2 (a weak bias towards small V).
~,~I~,A
~""~
,_
~,~I- - - -cj~
:L \ .~
\_
~,1:- - - - - - c.- - - -: :,fc- \~"'~\_ ~
~t ,. A. 1 ~,~
I -------c:L,.
-----=-----'
-----=-----,1
-----=-----'
A-----"-----..t
, /____,________=_\
Figure 1: Belief states. Left panel: overhead map of corridor with initial belief b1 ;
belief state bso (just before robot finds out which direction it's pointing); belief bgo
(just after finding out). Right panel: reconstruction of bso with 3, 4, and 5 features.
We set G* = 1O- 1211U11 2j2 +6..(U), where 6.. is 0 when the first column of U contains
all Is and 00 otherwise. This loss function fixes the first column of U, representing
our knowledge that one feature should be a normalizing constant so that each belief
sums to 1. The subgradient of G* is 1O- 12U + [k, 0], so equation (5) becomes
(X - exp(UV))VT
= 1O- 12U + [k, 0]
Here [k,O] is a matrix with an arbitrary first column and all other elements 0; this
matrix has enough degrees of freedom to compensate for the constraints on U.
Our Newton algorithm handles this modified fixed point equation without difficulty.
So, this (GL)2M is a principled and efficient way to decompose a matrix of probability distributions. So far as we know this model and algorithm have not been
described in the literature.
Figure 1 shows our reconstructions of a representative belief state using I = 3, 4,5
features (one of which is a normalizing constant that can be discarded for planning).
The I = 5 reconstruction is consistently good across all 400 beliefs, while the I = 4
reconstruction has minor artifacts for some beliefs. A small number of restarts is
required to achieve good decompositions for I = 3 where the optimization problem
is most constrained. For comparison, a traditional SVD requires a matrix of rank
about 25 to achieve the same mean-squared reconstruction error as our rank-3
decomposition. It requires rank about 85 to match our rank-5 decomposition.
Examination of the learned U matrix (not shown) for I = 4 reveals that the corridor is mapped into two smooth curves in feature space, one for each orientation.
Corresponding states with opposite orientations are mapped into similar feature
vectors for one half of the corridor (where the training beliefs were sometimes confused about orientation) but not the other (where there were no training beliefs
that indicated any connection between orientations). Reconstruction artifacts occur
when a curve nearly self-intersects and causes confusion between states. This selfintersection happens because of local minima in the loss function; with more flexibility (I = 5) the optimizer is able to untangle the curves and avoid self-intersection.
Our success in compressing the belief state translates directly into success in planning; see [15] for details. By comparison, traditional SVD on either the beliefs or
the log beliefs produces feature sets which are unusable for planning because they
do not achieve sufficiently good reconstruction with few enough features.
8
Discussion
We have introduced a new general class of nonlinear regression and factor analysis
model, presenting a derivation and algorithm, reductions to previously-known special cases, and a practical example. The model is a drop-in replacement for PCA,
but can provide much better reconstruction performance in cases where the PCA
error model is inaccurate. Future research includes online algorithms for parameter
adjustment; extensions for missing data; and exploration of new link functions.
Acknowledgments
Thanks to Nick Roy for helpful comments and for providing the data analyzed
in section 7. This work was supported by AFRL contract F30602-01-C-0219,
DARPA's MICA program, and by AFRL contract F30602- 98- 2- 0137, DARPA's
CoABS program. The opinions and conclusions are the author's and do not reflect
those of the US government or its agencies.
References
[1] T. K. Landauer, P . W . Foltz , and D. Laham . Introduction to latent semantic analysis .
Discourse Processes, 25:259- 284, 1998.
[2] Jon M. Kleinberg. Authoritative sources in a hyperlinked environment. Journal of
the ACM, 46(5) :604-632, 1999.
[3] M. Turk and A. Pentland. Eigenfaces for recognition. Journal of Cognitive Neuroscience, 3(1) :71-86, 1991.
[4] Carlo Tomasi and Takeo Kanade. Shape and motion from image streams under
orthography: a factorization method. Int. J. Computer Vision , 9(2):137- 154, 1992.
[5] D. P. O'Leary and S. Peleg. Digital image compression by outer product expansion.
IEEE Trans . Communications, 31:441-444, 1983.
[6] P . McCullagh and J. A. Neider. Generalized Linear Models. Chapman & Hall, London,
2nd edition, 1983.
[7] Peter Auer, Mark Hebster, and Manfred K. Warmuth. Exponentially many local
minima for single neurons. In NIPS, vol. 8. MIT Press, 1996.
[8] R. Tyrell Rockafellar. Convex Analysis. Princeton University Press, New Jersey, 1970.
[9] Geoffrey J. Gordon. Approximate Solutions to Markov Decision Processes.
thesis, Carnegie Mellon University, 1999.
PhD
[10] Daniel Lee and H. Sebastian Seung. Algorithms for nonnegative matrix factorization.
In NIPS, vol. 13. MIT Press, 2001.
[11] Nathan Srebro. Personal communication, 2002.
[12] Anthony J . Bell and Terrence J. Sejnowski. The 'independent components' of natural
scenes are edge filters. Vision Research, 37(23) :3327- 3338, 1997.
[13] Michael Collins, Sanjoy Dasgupta, and Robert Schapire. A generalization of principal
component analysis to the exponential family. In NIPS, vol. 14. MIT Press, 2002 .
[14] D. Fox, W. Burgard, F . Dellaert , and S. Thrun. Monte Carlo localization: Efficient
position estimation for mobile robots. In AAAI, 1999.
[15] Nicholas Roy and Geoffrey J. Gordon. Exponential family PCA for belief compression
in POMDPs. In NIPS, vol. 15 . MIT Press, 2003.
[16] Sam Roweis. EM algorithms for PCA and SPCA. In NIPS, vol. 10. MIT Press, 1998.
| 2144 |@word middle:1 compression:3 nd:1 open:1 seek:1 git:1 decomposition:5 simplifying:1 u11:1 pick:2 pressure:1 tr:1 reduction:1 initial:1 contains:4 zij:1 daniel:1 current:1 must:7 written:1 takeo:1 j1:1 shape:1 drop:2 update:4 half:1 guess:1 warmuth:1 ith:1 vanishing:1 lr:2 manfred:1 provides:1 simpler:1 along:4 become:1 corridor:10 combine:2 fitting:2 overhead:1 introduce:1 ica:4 expected:1 themselves:1 examine:1 planning:5 globally:1 resolve:1 becomes:4 confused:1 notation:3 panel:2 what:3 bso:2 finding:2 unobserved:1 guarantee:1 collecting:1 unit:1 positive:3 before:2 local:5 treat:1 tends:2 limit:2 bilinear:2 approximately:1 abuse:1 might:2 specifying:1 factorization:4 range:1 practical:1 acknowledgment:1 practice:1 implement:1 definite:1 procedure:2 zis:1 bell:1 matching:9 word:2 v11:1 get:1 applying:1 restriction:4 optimize:2 equivalent:1 map:3 missing:1 go:1 starting:1 convex:15 rectangular:1 factored:1 estimator:1 rule:1 pull:1 handle:1 traditionally:1 laham:1 updated:1 target:2 lighter:1 us:1 element:5 roy:2 recognition:2 particularly:1 updating:1 svds:1 ensures:1 compressing:1 decrease:1 mentioned:1 principled:1 environment:3 agency:1 ui:3 seung:1 personal:1 depend:1 solving:1 localization:1 darpa:2 jersey:1 intersects:1 derivation:1 describe:1 shortcoming:1 london:1 sejnowski:1 monte:1 tell:4 larger:1 solve:4 say:2 otherwise:1 favor:1 gi:2 think:1 online:1 differentiate:1 advantage:1 differentiable:2 hyperlinked:1 kurtosis:1 propose:2 reconstruction:10 interaction:1 product:3 j2:1 mixing:1 flexibility:1 achieve:4 roweis:1 getting:1 convergence:2 regularity:1 requirement:1 produce:2 help:1 derive:1 linearize:2 measured:1 ij:3 minor:1 c:1 predicted:1 implies:1 peleg:1 convention:1 direction:3 unew:1 filter:1 pea:2 exploration:1 duals:2 opinion:1 linear2:1 require:3 government:1 feeding:1 fix:2 generalization:3 wall:1 decompose:2 tyrell:1 multilinear:1 extension:1 hold:1 sufficiently:2 around:1 tracker:1 hall:1 exp:2 pointing:2 substituting:1 driving:1 optimizer:1 smallest:1 estimation:1 applicable:1 tanh:3 tool:2 weighted:3 mit:5 sensor:2 gaussian:2 always:3 modified:1 avoid:2 hj:3 mobile:1 focus:1 consistently:1 rank:7 bernoulli:1 check:2 likelihood:1 sense:1 helpful:1 inaccurate:2 hidden:1 reproduce:1 overall:4 dual:1 orientation:5 priori:2 plan:2 constrained:1 special:9 equal:1 never:1 chapman:1 represents:2 nearly:2 jon:1 future:1 lrd:2 di1:1 others:1 gordon:3 simplify:1 duplicate:1 few:2 dg:2 simultaneously:2 divergence:6 recognize:1 individual:1 replacement:2 maintain:1 freedom:1 analyzed:1 accurate:2 bregman:3 fu:1 edge:1 necessary:3 respective:1 orthogonal:1 unless:1 fox:1 increased:1 column:9 untangle:1 burgard:1 answer:2 thanks:1 contract:2 lee:1 terrence:1 infomax:1 picking:1 michael:1 leary:1 squared:8 reflect:1 thesis:1 aaai:1 reconstructs:1 choose:2 possibly:1 cognitive:1 inject:1 derivative:13 includes:1 int:1 rockafellar:1 satisfy:1 vi:3 stream:1 try:2 lab:1 neider:1 sort:1 recover:1 utx:1 square:3 accuracy:1 yield:2 identify:1 weak:1 accurately:1 carlo:2 multiplying:1 pomdps:1 whenever:1 sebastian:1 definition:2 hare:1 turk:1 di:3 knowledge:2 ut:5 cj:1 auer:1 worry:2 afrl:2 follow:2 restarts:1 specify:1 formulation:1 though:1 just:2 until:1 glms:1 hand:1 nonlinear:6 logistic:1 artifact:2 indicated:1 building:1 contain:1 alternating:1 iteratively:1 semantic:2 deal:1 reweighted:1 self:2 uniquely:1 unnormalized:2 m:11 generalized:11 linearizing:1 trying:1 presenting:1 confusion:1 motion:2 image:2 sigmoid:1 exponentially:2 tail:2 mellon:1 ai:1 uv:24 similarly:2 dj:1 dot:1 moving:1 robot:8 optimizes:1 success:3 arbitrarily:1 vt:3 yi:8 minimum:5 additional:2 greater:1 lfij:1 fortunately:1 determine:1 redundant:1 relates:1 multiple:1 reduces:2 smooth:1 match:3 faster:1 long:2 retrieval:1 compensate:1 prediction:11 involving:1 regression:4 essentially:1 cmu:1 df:2 poisson:2 vision:2 iteration:1 sometimes:1 orthography:1 addition:1 want:1 interval:1 singular:1 grow:1 source:2 comment:1 tend:1 call:2 near:2 door:1 spca:1 enough:3 fit:1 zi:6 restrict:1 opposite:1 reduce:3 mica:1 translates:1 cousin:1 det:1 whether:1 expression:2 pca:7 peter:1 hessian:3 cause:1 dellaert:1 useful:2 proportionally:1 listed:1 cosh:2 simplest:1 reduced:4 generate:1 schapire:1 exist:1 neuroscience:1 write:4 carnegie:1 dasgupta:1 vol:5 group:1 redundancy:1 subgradient:1 sum:6 inverse:1 uncertainty:1 place:1 family:5 reasonable:1 saying:1 dj1:1 decision:1 entirely:1 hi:1 quadratic:1 nonnegative:2 occur:1 constraint:1 scene:1 kleinberg:1 nathan:1 simulate:1 argument:3 span:2 subgradients:2 smaller:1 across:1 em:1 sam:1 modification:1 happens:2 equation:13 previously:1 turn:1 know:2 tractable:1 end:3 apply:1 nicholas:1 encounter:2 original:2 assumes:1 remaining:1 include:5 newton:7 f30602:2 diagonal:1 traditional:2 div:1 gradient:4 navigating:1 subspace:2 link:30 unable:1 thrun:1 separate:2 majority:2 sensible:1 nondifferentiable:1 lateral:1 mapped:2 outer:1 collected:2 providing:1 minimizing:3 equivalently:1 difficult:1 unfortunately:1 robert:1 potentially:1 info:2 negative:1 implementation:1 unknown:2 allowing:1 neuron:1 markov:1 discarded:1 descent:3 pentland:1 communication:2 arbitrary:1 introduced:1 pair:2 required:1 kl:2 componentwise:1 connection:1 nick:1 tomasi:1 learned:1 emu:1 nip:5 trans:1 address:1 able:2 below:4 pattern:1 usually:2 reading:1 program:2 including:2 max:2 belief:27 power:1 natural:2 difficulty:1 examination:1 predicting:1 representing:2 extract:1 bare:1 speeding:1 prior:2 literature:1 loss:38 interesting:1 analogy:1 geoffrey:3 facing:1 srebro:1 digital:1 authoritative:1 degree:1 imposes:1 vij:1 heavy:3 row:4 lo:1 course:1 gl:18 supported:1 copy:2 jth:1 side:1 bias:1 understand:1 foltz:1 eigenfaces:2 curve:3 dimension:3 valid:1 world:1 doesn:1 stuck:1 author:1 far:2 approximate:1 observable:1 ignore:1 keep:2 reveals:1 b1:1 xi:3 landauer:1 don:2 search:1 iterative:1 ggordon:1 latent:2 decomposes:1 vergence:1 tailed:2 kanade:1 ignoring:1 expansion:1 artificially:1 anthony:1 domain:1 vj:3 noise:5 edition:1 representative:1 vr:1 precision:1 position:4 exponential:5 minz:1 ito:1 unusable:1 bad:1 xt:1 list:1 normalizing:2 exists:2 vnew:1 phd:1 easier:1 aren:1 intersection:1 logarithmic:1 fc:1 likely:2 adjustment:1 corresponds:1 discourse:1 dh:2 acm:1 goal:2 identity:2 towards:2 replace:3 mccullagh:1 justify:1 principal:5 called:1 sanjoy:1 duality:1 e:1 svd:9 mark:1 latter:1 collins:1 princeton:1 |
1,257 | 2,145 | Learning Sparse Multiscale Image
Representations
Phil Sallee
Department of Computer Science and
Center for Neuroscience, UC Davis
1544 Newton Ct.
Davis, CA 95616
[email protected]
Bruno A. Olshausen
Department of Psychology and
Center for Neuroscience, UC Davis
1544 Newton Ct.
Davis, CA 95616
[email protected]
Abstract
We describe a method for learning sparse multiscale image representations using a sparse prior distribution over the basis function
coefficients. The prior consists of a mixture of a Gaussian and a
Dirac delta function, and thus encourages coefficients to have exact
zero values. Coefficients for an image are computed by sampling
from the resulting posterior distribution with a Gibbs sampler. The
learned basis is similar to the Steerable Pyramid basis, and yields
slightly higher SNR for the same number of active coefficients. Denoising using the learned image model is demonstrated for some
standard test images, with results that compare favorably with
other denoising methods.
1
Introduction
Increasing interest has been given to the use of overcomplete representations for
natural scenes, where the number of basis functions exceeds the number of image
pixels. One reason for this is that overcompleteness allows for more stable, and thus
arguably more meaningful, representations in which common image features can be
well described by only a few coefficients, regardless of where they are located in
the image, how they are rotated, or how large they are [8, 6]. This may translate
into gains in coding efficiency for image compression, and improved accuracy for
tasks such as denoising. Overcomplete representations have been shown to reduce
Gibbs-like artifacts common to thresholding methods employing critically sampled
wavelets [4, 3, 9].
Common wavelet denoising approaches generally apply either a hard or softthresholding function to coefficients which have been obtained by filtering an image
with a the basis functions. One can view these thresholding methods as a means
of selecting coefficients for an image based on an assumed sparse prior on the coefficients [1, 2]. This statistical framework provides a principled means of selecting
an appropriate thresholding function. When such thresholding methods are applied
to overcomplete representations, however, problems arise due to the dependencies
between coefficients. Choosing optimal thresholds for a non-orthogonal basis is still
an unsolved problem. In one approach, orthogonal subgroups of an overcomplete
shift-invariant expansion are thresholded separately and then the results are combined by averaging [4, 3]. In addition, if the coefficients are obtained by filtering
the noisy image, there will be correlations in the noise that should be taken into
account.
Here we address two major issues regarding the use of overcomplete representations
for images. First, current methods make use of various overcomplete wavelet bases.
What is the optimal basis to use for a specific class of data? To help answer this
question, we describe how to adapt an overcomplete wavelet basis to the statistics of natural images. Secondly, we address the problem of properly inferring the
coefficients for an image when the basis is overcomplete. We avoid problems associated with thresholding by using the wavelet basis as part of a generative model,
rather than a simple filtering mechanism. We then sample the coefficients from
the resulting posterior distribution by simulating a Markov process known as a
Gibbs-sampler.
Our previous work in this area made use of a prior distribution peaked at zero and
tapering away smoothly to obtain sparse coefficients [7]. However, we encountered
a number of significant limitations with this method. First, the smooth priors
do not force inactive coefficients to have values exactly equal to zero, resulting in
decreased coding efficiency. Efficiency may be partially regained by thresholding
the near-zero coefficients, but due to the non-orthogonality of the representation
this will produce sub-optimal results as previously mentioned. The maximum a
posteriori (MAP) estimate also introduced biases in the learning process. These
effects can be partially compensated for by renormalizing the basis functions, but
other parameters of the model such as those of the prior could not be learned.
Finally, the gradient ascent method has convergence problems due to the power
spectrum of natural images and the overcompleteness of the representation. Here we
resolve these problems by using a prior distribution which is composed of a mixture
of a Gaussian and a Dirac delta function, so that inactive coefficients are encouraged
to have exact zero values. Similar models employing a mixture of two Gaussians
have been used for classifying wavelet coefficients into active (high variance) and
inactive (low variance) states [2, 5]. Such a classification should be even more
advantageous if the basis is overcomplete. A method for performing Gibbs-sampling
for the Delta-plus-Gaussian prior in the context of an image pyramid is derived, and
demonstrated to be effective at obtaining very sparse representations which match
the form of the imposed prior. Biases in the learning are overcome by sampling
instead of using a MAP estimate.
2
Wavelet image model
Each observed image I is assumed to be generated by a linear superposition of basis
functions which are columns of an N by M weight matrix W, with the addition of
Gaussian noise ?:
I = W a + ?,
(1)
where I is an N -element vector of image pixels and a is an M -element vector of basis
coefficients. In order to achieve a practical implementation which can be seamlessly
scaled to any size image, we assume that the basis function matrix W is composed of
a small set of spatially localized mother wavelet functions ?i (x, y), which are shifted
to each position in the image and rescaled by factors of two. Unlike typical wavelet
transforms which use a single 1-D mother wavelet function to generate 2-D functions
by inner product, we do not constrain the functions ?i (x, y) to be 1-D separable.
The functions ?i (x, y) provide an efficient way to perform computations involving
W by means of convolutions. Basis functions of coarser scales are produced by
upsampling the ?i (x, y) functions and blurring with a low-pass filter ?(x, y), also
known as the scaling function. The image model above may be re-expressed to
make these parameters explicit:
I(x, y)
g l (x, y)
= g 0 (x, y) + ?(x, y)
l+1
P
g (x, y) ? 2 ? ?(x, y) + i ali (x, y) ? ?i (x, y)
=
al (x, y)
(2)
l <L?1
(3)
l =L?1
where the coefficients ali (x, y) are indexed by their position (x, y), band (i) and
level of resolution (l) within the pyramid (l = 0 is the highest resolution level). The
symbol ? denotes convolution, and ? 2 denotes upsampling by two and is defined as
f ( x2 , y2 ) x even & y even
f (x, y) ? 2 ?
(4)
0
otherwise
The probability of generating an image I, given coefficients a, parameters ?, assuming Gaussian i.i.d. noise ? (with variance 1/?N ), is
P (I|a, ?)
1 ? ?N |I?W a|2
.
e 2
Z ?N
=
(5)
The prior probability over each coefficient ai is modeled as a mixture of a Gaussian
distribution and a Dirac delta function ?(ai ). A binary state variable si for each
coefficient indicates whether the coefficient ai is active (any real value), or inactive
(zero). The probability of a coefficient vector a given a binary state vector s and
model parameters ? = {W, ?N , ?a , ?s } is defined as
Y
P (ai |si , ?)
(6)
P (a|s, ?) =
i
P (ai |si , ?)
=
(
?(ai )
1
Z?a
e
?
?a
i
2
a2i
if
si = 0,
if
si = 1
(7)
i
where ?a is a vector with elements ?ai . The probability of a binary state s is
P (s|?) =
1 ? 1 sT ? s s
.
e 2
Z ?s
(8)
Matrix ?s is assumed to be diagonal (for now), with nonzero elements ?si . The
form of the prior is shown graphically in figure 1. Note that the parameters W, ? a ,
and ?s are themselves parameterized by a much smaller set of parameters. Only
the mother wavelet function ?i (x, y) and a single ?si and ?ai parameter need to be
learned for each wavelet band, since we are assuming translation invariance.
The total image probability is obtained by marginalizing over the possible coefficient
and state values:
Z
X
P (I|?) =
P (s|?) P (I|a, ?)P (a|s, ?) da
(9)
s
3
Sampling and Inference
We show how to sample from the posterior distribution P (a, s|I, ?) for an image
I using a Gibbs sampler. For each coefficient and state variable pair (ai ,si ), we
10
0
-1
10
-2
10
-3
10
-4
10
-5
10
-6
10
0.5
0.4
0.3
0.2
0.1
0
0.1
0.2
0.3
0.4
0.5
Figure 1: Prior distribution (dashed), and histogram of samples taken from the
posterior distribution (solid) plotted for a single coefficient. The y-axis is plotted
on a log scale.
sample from the posterior distribution conditioned on the image and the remaining
coefficients a?i : P (ai , si |I, a?i , s?i , ?). After all coefficients (and state variables) have
been updated, this process is repeated until the system has reached equilibrium. To
infer an optimal representation a for an image I (for coding or denoising purposes),
we can either average a number of samples to estimate the posterior mean, or with
minor adjustment locate a posterior maximum by raising the posterior distribution
to a power (1/T ) and annealing T to zero. To sample from P (ai , si |I, a?i , s?i , ?), we
first draw a value for si from P (si |I, a?i , s?i , ?), then draw ai from P (ai |si , I, a?i , s?i , ?).
For P (si |I, a?i , s?i , ?) we have:
P (si |I, a?i , s?i , ?)
? P (si |s?i , ?)
where
P (si |s?i , ?)
=
P (I|ai , a?i , ?)
=
Z
P (I|ai , a?i , ?)P (ai |si , ?)dai
1
e?
si
,
Zsi |s?i
1 ? ?ni (ai ?bi )2
e 2
,
Z ?ni
and
?ni = ?N |Wi |2 ,
?s
i
2
bi =
Wi ? (I ? W ai=0 )
.
|Wi |2
(10)
(11)
(12)
(13)
The notation Wi denotes column i of matrix W, |Wi | is the length of vector Wi ,
and ai=0 denotes the current coefficient vector a except with ai set to zero. Thus,
bi denotes the value for ai which minimizes the reconstruction error (while holding
a?i constant). Since si can only take on two values, we can compute equation 10 for
si = 0 and si = 1, integrating over the possible coefficient values. This yields the
following sigmoidal activation rule as a function of bi :
P (si = 1|I, a?i , s?i , ?)
=
1
1+e
??i (b2i ?ti )
(14)
where
?2ni
1
?i =
,
2 ? n i + ? ai
? n i + ? ai
? ai
ti =
?si ? log
.
?2ni
? n i + ? ai
(15)
For P (ai |si , I, a?i , s?i , ?) we have:
P (ai |si , I, a?i , s?i , ?)
=
(
?(ai )
? i bi
, ?n
N ( ?n n+?
a
i
i
1
+?ai
i
)
if si = 0,
if si = 1
(16)
To perform this procedure on a wavelet pyramid, the inner product computations
necessary to compute bi can be performed efficiently by means of convolutions with
the mother wavelet functions ?i (x, y). The ?N , ?si and ?ai parameters may be
adapted to a specific image during the inference process by use of the update rules
described in the next section. This method was found to be particularly useful for
denoising, when the variance of the noise was assumed to be unknown.
4
Learning
Our objective for learning is to adjust the parameters, ?, to maximize the average
log-likelihood of images under the model:
?? = arg max hlog P (I|?)i
(17)
?
The parameters are updated by gradient ascent on this objective, which results in
the following update rules:
??si
??ai
??i (x, y)
?
?
1
2
*
1
2
*
1
1
1 + e 2 ? si
si
? si
1
? a2i
? ai
P (a,s|I,?)
P (a,s|I,?)
+
D
E
? ?N he(x, y) ? ai (x, y)iP (a,s|I,?)
+
(18)
(19)
(20)
where ? denotes cross correlation and e(x, y) is the reconstruction error computed
by e = I ? W a. Only a center portion of the cross correlation with the extent of
the ?i (x, y) functions is computed to update the parameters. The outer brackets
denotes averaging over many images. The notation hiP () denotes averaging the
quantity in brackets while sampling from the specified distribution.
5
Results
The image model was trained on 22 512x512 pixel grayscale natural images (not
whitened). These images were generated from color images taken from a larger
database of photographic images 1 . Smaller images (64x64 pixels) were selected
randomly for sampling during training. To simplify the learning procedure, sampling was performed on a single spatial frequency scale. Each image was bandpass
filtered for an octave range before sampling from the posterior for that scale. The
1
Images were downloaded from philip.greenspun.com with permission from Philip
Greenspun.
(a)
(b)
Figure 2: (a) Mother wavelet functions ?i (x, y) adapted for 2, 4 and 6 bands and
corresponding power spectra showing power as a function of spatial frequency in
the 2D Fourier plane. (b) Equivalent mother wavelets and spectra for the 4-band
Steerable Pyramid.
?ai and ?si parameters were constrained to be the same for all orientation bands and
were adapted over many images with ?N fixed. Shown in figure 2 are the learned
?i (x, y) which parameterize W , with their corresponding 2D spectra. Three different degrees of overcompleteness were tested. The results are shown for 2 band, 4
band and 6 band wavelet bases. As the degree of overcompleteness increases, the
resulting functions show tighter tuning to orientation. The basis filters for a 4 band
Steerable Pyramid [10] are also shown for comparison, to illustrate the similarity
to the learned functions.
27
learned
steer
26.5
26
SNR (dB)
25.5
25
24.5
24
23.5
23
22.5
1.0
2.0
3.0
4.0
5.0
% nonzeros
Figure 3: Sparsity comparison between the learned basis (top) and the steerable
basis (bottom). The y axis represents the signal-to-noise ratio (SNR) in dB achieved
for each method for a given percentage of nonzeros.
5.1
Sparsity
We evaluated the sparsity of the representations obtained with the four band learned
functions and the sampling method with those obtained using the same sampling
method and the four band Steerable Pyramid filters [10]. In order to explore the
SNR curves for each basis, we used a variety of values for ?s so as to obtain different
levels of sparsity. The same images were used for both bases. The results are given
in figure 3. Each dot on the line represents a different value of ?s . The results were
similar, with the learned basis yielding slightly higher SNR (about 0.5 dB) for the
same number of active coefficients.
5.2
Denoising
We evaluated our inference method and learned basis functions by denoising images
containing known amounts of additive i.i.d. Gaussian noise. Denoising was accomplished by averaging samples taken from the posterior distribution for each image
via Gibbs sampling to approximate the posterior mean. Gibbs sampling was performed on a four level pyramid using the 6 band learned wavelet basis, and also
using the 6 band Steerable basis. The ?N , ?si and ?ai parameters were adapted to
each noisy image during sampling for blind denoising in which the noise variance
was assumed to be unknown. We compared these results to the wiener2 function in
MATLAB, and also to BayesCore [9], a Bayesian method for computing an optimal
soft thresholding, or coring, function for a generalized Laplacian prior. For wiener2,
the best neighborhood size was used for each image. Table 1 gives the SNR results
for each method when applied to some standard test images for three different levels of i.i.d. Gaussian noise with standard deviation ?. Figure 4 shows a cropped
subregion of the results for the ?Einstein? image with ? = 10.
6
Summary and Conclusions
We have shown that a wavelet basis and a mixture prior composed of a Dirac delta
function and a Gaussian can be adapted to natural images resulting in very sparse
image representations. The resulting basis is very similar to a Steerable basis, both
in appearance and sparsity of the resulting image representations. It appears that
the Steerable basis may be nearly optimal for producing sparse representations of
natural scenes. Denoising results indicate that using a sparse prior and an inference
method to properly account for the non-orthogonality of the representation may
yield a significant improvement over wavelet coring methods that use filtered coefficients. More work needs to be done to determine whether the coding gains achieved
are due to the choice of prior versus the basis or inference/estimation method used.
Acknowledgments Supported by NIMH R29-MH057921. Phil Sallee?s work was
also supported in part by a United States Department of Education Government
Assistance in Areas of National Need (DOE-GAANN) grant #P200A980307.
Image
Einstein
Lena
Goldhill
Fruit
noise level
? = 10
? = 20
? = 30
? = 10
? = 20
? = 30
? = 10
? = 20
? = 30
? = 10
? = 20
? = 30
noisy
12.40
6.40
2.89
13.61
7.59
4.07
13.86
7.83
4.28
16.25
10.24
6.70
wiener2
15.80
12.61
10.95
19.05
15.51
13.25
17.56
14.32
12.64
21.87
18.15
15.97
BayesCore S6
16.36
13.44
11.81
19.91
16.88
14.99
18.14
15.18
13.61
22.09
18.97
17.21
D+G S6
16.47
13.80
12.28
20.37
17.46
15.48
18.10
15.41
13.92
22.78
19.61
17.72
D+G L6
16.19
13.79
12.29
20.21
17.54
15.55
17.90
15.41
13.95
22.38
19.42
17.66
Table 1: SNR values (in dB) for noisy and denoised images contaminated with
additive i.i.d. Gaussian noise of std.dev. ?. ?D+G? means Delta-plus-Gaussian
prior, ?S6? means 6-Band Steerable basis, and ?L6? means 6-Band Learned basis.
original
noisy (?=10) SNR=12.3983
wiener2 SNR=15.8033
BayesCore steer6 SNR=16.3591
D+G steer6 SNR=16.4714
D+G learned6 SNR=16.1939
Figure 4: Denoising example. A cropped subregion of the Einstein image and
denoised images for each noise reduction method for noise std.dev. ?=10.
References
[1] Abromovich F, Sapatinas T, Silverman B (1996), Wavelet Thresholding via a Bayesian
Approach, preprint.
[2] Chipman H, Kolaczyk E, McCulloch R (1997) Adaptive bayesian wavelet shrinkage,
J. Amer. Statist. Assoc. 92(440): 1413-1421.
[3] Chang SG, Yu B, Vetterli M (2000). Spatially Adaptive Wavelet Thresholding with
Context Modelling for Image Denoising. IEEE Trans. on Image Proc., 9(9): 1522-1531.
[4] Coifman RR, Donoho DL (1995). Translation-invariant de-noising, in Wavelets and
Statistics, A.Antoniadis and G. Oppenheim, Eds. Berlin, Germany: Springer-Varlag.
[5] Crouse MS, Nowak RD and Baraniuk RG (1998) Wavelet-based Statistical Signal
Processing using Hidden Markov Models, IEEE Trans. Signal Proc., 46(4): 886-902.
[6] Freeman WT, Adelson EH (1991) The Design and Use of Steerable Filters. IEEE
Trans. Patt. Anal. and Machine Intell., 13(9): 891-906.
[7] Olshausen BA, Sallee P, Lewicki MS (2001) Learning sparse image codes using a
wavelet pyramid architecture, Adv. in Neural Inf. Proc. Sys., 13: 887-893.
[8] Simoncelli EP, Freeman WT, Adelson EH, Heeger DJ (1992) Shiftable multiscale transforms, IEEE Transactions on Information Theory, 38(2): 587-607.
[9] Simoncelli EP, Adelson EH (1996) Noise removal via Bayesian wavelet coring, Presented at: 3rd IEEE International Conf. on Image Proc., Laussanne Switzerland.
[10] Simoncelli EP, Freeman WT (1995). The Steerable Pyramid: A Flexible Architecture
for Multi-scale Derivative Computation, IEEE Int. Conf. on Image Processing.
| 2145 |@word kolaczyk:1 compression:1 advantageous:1 solid:1 reduction:1 tapering:1 selecting:2 united:1 current:2 com:1 si:35 activation:1 additive:2 update:3 generative:1 selected:1 antoniadis:1 plane:1 sys:1 filtered:2 provides:1 sigmoidal:1 consists:1 coifman:1 wiener2:4 themselves:1 multi:1 lena:1 freeman:3 resolve:1 increasing:1 notation:2 mcculloch:1 what:1 minimizes:1 ti:2 exactly:1 scaled:1 assoc:1 grant:1 producing:1 arguably:1 before:1 plus:2 bi:6 range:1 practical:1 acknowledgment:1 silverman:1 x512:1 procedure:2 steerable:11 area:2 integrating:1 noising:1 context:2 equivalent:1 map:2 demonstrated:2 center:3 phil:2 compensated:1 graphically:1 regardless:1 imposed:1 resolution:2 rule:3 s6:3 x64:1 updated:2 exact:2 element:4 particularly:1 located:1 std:2 coarser:1 database:1 observed:1 bottom:1 ep:3 preprint:1 parameterize:1 adv:1 rescaled:1 highest:1 principled:1 mentioned:1 nimh:1 trained:1 ali:2 efficiency:3 blurring:1 basis:31 various:1 describe:2 effective:1 choosing:1 neighborhood:1 larger:1 otherwise:1 statistic:2 noisy:5 ip:1 rr:1 reconstruction:2 product:2 translate:1 achieve:1 dirac:4 convergence:1 produce:1 renormalizing:1 generating:1 rotated:1 help:1 illustrate:1 minor:1 subregion:2 c:1 indicate:1 switzerland:1 filter:4 education:1 government:1 tighter:1 secondly:1 equilibrium:1 major:1 purpose:1 estimation:1 proc:4 superposition:1 overcompleteness:4 gaussian:11 rather:1 avoid:1 shrinkage:1 derived:1 properly:2 improvement:1 modelling:1 indicates:1 likelihood:1 seamlessly:1 laussanne:1 posteriori:1 inference:5 hidden:1 germany:1 pixel:4 issue:1 classification:1 arg:1 orientation:2 flexible:1 spatial:2 constrained:1 uc:2 equal:1 sampling:13 encouraged:1 represents:2 yu:1 adelson:3 nearly:1 peaked:1 contaminated:1 simplify:1 few:1 randomly:1 composed:3 national:1 intell:1 interest:1 baolshausen:1 adjust:1 mixture:5 bracket:2 yielding:1 nowak:1 necessary:1 orthogonal:2 indexed:1 re:1 plotted:2 overcomplete:9 hip:1 column:2 soft:1 steer:1 dev:2 deviation:1 shiftable:1 sallee:4 snr:12 dependency:1 answer:1 combined:1 st:1 international:1 containing:1 conf:2 derivative:1 account:2 de:1 coding:4 coefficient:33 int:1 blind:1 b2i:1 performed:3 view:1 reached:1 portion:1 denoised:2 ni:5 accuracy:1 variance:5 efficiently:1 yield:3 bayesian:4 critically:1 produced:1 zsi:1 oppenheim:1 ed:1 frequency:2 associated:1 unsolved:1 gain:2 sampled:1 color:1 vetterli:1 appears:1 higher:2 improved:1 amer:1 evaluated:2 done:1 correlation:3 until:1 multiscale:3 artifact:1 olshausen:2 effect:1 y2:1 spatially:2 nonzero:1 assistance:1 during:3 encourages:1 davis:4 m:2 generalized:1 octave:1 image:59 common:3 he:1 significant:2 gibbs:7 mother:6 ai:35 tuning:1 rd:2 bruno:1 dj:1 dot:1 stable:1 similarity:1 base:3 posterior:11 inf:1 binary:3 accomplished:1 regained:1 dai:1 goldhill:1 determine:1 maximize:1 dashed:1 signal:3 photographic:1 simoncelli:3 nonzeros:2 infer:1 exceeds:1 smooth:1 match:1 adapt:1 cross:2 laplacian:1 involving:1 whitened:1 histogram:1 pyramid:10 achieved:2 addition:2 cropped:2 separately:1 decreased:1 annealing:1 unlike:1 ascent:2 db:4 chipman:1 near:1 variety:1 psychology:1 architecture:2 reduce:1 regarding:1 inner:2 shift:1 inactive:4 whether:2 matlab:1 generally:1 useful:1 transforms:2 amount:1 band:15 statist:1 generate:1 percentage:1 shifted:1 neuroscience:2 delta:6 patt:1 four:3 threshold:1 thresholded:1 parameterized:1 baraniuk:1 draw:2 scaling:1 ct:2 encountered:1 adapted:5 orthogonality:2 constrain:1 scene:2 x2:1 fourier:1 performing:1 separable:1 department:3 smaller:2 slightly:2 wi:6 invariant:2 taken:4 equation:1 previously:1 mechanism:1 gaussians:1 mh057921:1 apply:1 einstein:3 away:1 appropriate:1 simulating:1 permission:1 softthresholding:1 a2i:2 original:1 denotes:8 remaining:1 top:1 newton:2 l6:2 objective:2 question:1 quantity:1 diagonal:1 gradient:2 berlin:1 upsampling:2 philip:2 outer:1 extent:1 reason:1 assuming:2 length:1 code:1 modeled:1 ratio:1 hlog:1 holding:1 favorably:1 ba:1 implementation:1 design:1 anal:1 unknown:2 perform:2 convolution:3 markov:2 locate:1 introduced:1 pair:1 specified:1 raising:1 learned:13 ucdavis:2 subgroup:1 trans:3 address:2 sparsity:5 max:1 power:4 natural:6 force:1 eh:3 axis:2 r29:1 prior:17 sg:1 removal:1 marginalizing:1 limitation:1 filtering:3 versus:1 localized:1 downloaded:1 degree:2 coring:3 fruit:1 thresholding:9 classifying:1 translation:2 summary:1 supported:2 bias:2 sparse:10 overcome:1 curve:1 made:1 adaptive:2 employing:2 transaction:1 approximate:1 active:4 assumed:5 spectrum:4 grayscale:1 table:2 ca:2 obtaining:1 expansion:1 da:1 noise:13 arise:1 repeated:1 sub:1 inferring:1 position:2 explicit:1 bandpass:1 heeger:1 wavelet:27 specific:2 showing:1 symbol:1 dl:1 conditioned:1 smoothly:1 rg:1 explore:1 appearance:1 expressed:1 adjustment:1 partially:2 lewicki:1 chang:1 springer:1 donoho:1 hard:1 typical:1 except:1 sampler:3 averaging:4 denoising:13 wt:3 total:1 pas:1 invariance:1 meaningful:1 tested:1 |
1,258 | 2,146 | Conditional Models on the Ranking Poset
Guy Lebanon
School of Computer Science
Carnegie Mellon University
Pittsburgh, PA 15213
John Lafferty
School of Computer Science
Carnegie Mellon University
Pittsburgh, PA 15213
[email protected]
[email protected]
Abstract
A distance-based conditional model on the ranking poset is presented
for use in classification and ranking. The model is an extension of the
Mallows model, and generalizes the classifier combination methods
used by several ensemble learning algorithms, including error correcting
output codes, discrete AdaBoost, logistic regression and cranking. The
algebraic structure of the ranking poset leads to a simple Bayesian interpretation of the conditional model and its special cases. In addition to a
unifying view, the framework suggests a probabilistic interpretation for
error correcting output codes and an extension beyond the binary coding
scheme.
1 Introduction
Classification is the task of associating a single label
with a covariate . A generalization of this problem is conditional ranking, the task of assigning to a full or partial
ranking of the items in . This paper studies the algebraic structure of this problem, and
proposes a combinatorial structure called the ranking poset for building probability models
for conditional ranking.
In ensemble approaches to classification and ranking, several base models are combined to
produce a single ranker or classifier. An important distinction between different ensemble
methods is whether they use discrete inputs, ranked inputs, or confidence-rated predictions. In the case of discrete inputs, the base models provide a single item in , and no
preference for a second or third choice is given. In the case of ranked input, the base classifiers output a full or partial ranking over . Of course, discrete input is a special case
of ranked input, where the partial ranking consists of the single topmost item. In the case
of confidence-rated predictions, the base models again output full or partial rankings, but
in addition provide a confidence score, indicating how much one class should be preferred
to another. While confidence-rated predictions are sometimes preferable as input to an ensemble method, such confidence scores are often not available (as is typically the case in
metasearch), and even when they are available, the scores may not be well calibrated.
This paper investigates a unifying algebraic framework for ensemble methods for classification and conditional ranking, focusing on the cases of discrete and ranked inputs.
, which consists of
Our approach is based on the ranking poset on items, denoted
the collection of all full and partial rankings equipped with the partial order given by re-
finement of rankings. The structure of the poset of partial ranking over gives rise to
natural invariant distance functions that generalize Kendall?s Tau and the Hamming distance. Using these distance functions we define a conditional model
where
. This conditional model generalizes several existing models
for classification and ranking, and includes as a special case the Mallows model [11]. In
addition, the model represents algebraically the way in which input classifiers are combined
in certain ensemble methods, including error correcting output codes [4], several versions
of AdaBoost [7, 1], and cranking [10].
In Section 2 we review some basic algebraic concepts and in Section 3 we define the ranking poset. The new model and its Bayesian interpretation are described in Section 4. A
derivation of some special cases is given in Section 5, and we conclude with a summary in
Section 6.
2 Permutations and Cosets
We begin by reviewing some basic concepts from algebra, with some of the notation and
definitions borrowed from Critchlow [2].
Identifying the items to be ranked
with the numbers
, if denotes a
permutation of
, then !"
denotes the rank given to item " and $# "
denotes
the item assigned to rank " . The collection of all permutations of -items forms the nonabelian symmetric group of order , denoted % . The multiplicative notation '& )(*
is used to denote function composition.
#
The subgroup of
; thus,
%
%
consisting of all permutations that fix the top
%
,(- .
%
#
The right coset
#
%
+ positions is denoted
!"
!(/"0
" (1
+2
34(53 67 '
%
#
(1)
8
(2)
is equivalent to a partial ranking, where there is a full ordering of the + top-ranked items.
.
The set of all such partial rankings forms the quotient space % :9%
#
= of positive integers that sum
An ordered partition of is a sequence ;<(
to . Such an ordered partition corresponds to a partial ranking of type ; with items
in the first position, > items in the second position and so on. No further information
is conveyed about orderings within each position. A partial ranking of the top + items is
aL special case with ?@
L (A+CB*
DE(F ( @(FG
L IHDE( .JK+ . More formally, let
(*GG
G
>M(N BOG
B >G
&&&
= (* BO&&&PB = # BO
.
Then the subgroupLR%R[ QOS( LR
% [ GTVUW&&&2UX% ZY contains all permutations W% for which
!
\
(
the set equality
holds for each " ; that is, all permutations that only permute
LR[
within each . A partial ranking of type ; is equivalent to a coset % Q and the set of such
partial rankings forms the quotient space % :9%]Q .
We now describe a convenient notation for permutations and cosets. In the following, we
list items separated by vertical lines, indicating that the items on the left side of the line are
preferred to (ranked higher than) the items on the right side of the line. For example, the
permutation !73M(*^
7!_^G`(a
!cbd(*b is denoted by ^:e b . A partial ranking %gf
#h
where the top 3 items are b
^
is denoted by b2 ^i8 j
k . A classification
may thus be
h
denoted by be
^
7j
k . A partial ranking % Q where ;l(mb:
^ with items
b
k ranked in
the first position is denoted by G
b:
k: ^n
j .
A distance function o on %
is a function oNpq% rUl% Nsut that satisfies the usual
properties: o2
7g(wv , o2
xyv when a(A
z , o2
C(wo2c
, and the triangle
T
PSfrag replacements
T
T
T
T
T
T
T
T
T
T
PSfrag replacements
T
Figure 1: The Hasse diagram of
h (left) and a partial Hasse diagram of (right). Some
of the lines are dotted for easier visualization.
o2
:B o2
for all
.% . In addition, since the indexing
inequality oc
of the items 8
is arbitrary, it is appropriate to require invariance to relabeling of .
Formally, this amounts to right invariance o2
(/o2:
: , for all
'%
A popular right invariant distance on %
]c
(
#
[
is Kendall?s Tau ]
, given by
[ # c "
JX # _7
(3)
where 2E( for ,xAv and 2E(4v otherwise [8]. Kendall?s Tau ]
can be
interpreted as the number of discordant pairs of items between and , or the minimum
number of adjacent transpositions needed to bring $# to D# . An adjacent transposition
flips a pair of items that have adjacent ranks. Critchlow [2] derives extensions of Kendall?s
Tau and other distances on %
to distances on partial rankings.
3 The Ranking Poset
We first define partially ordered sets and then proceed to define the ranking poset. Some of
the definitions below are taken from [12], where a thorough introduction to posets can be
found.
A partially ordered set or poset is a pair "!d
$#q , where ! is a set and # is a binary relation
that satisfies (1) %#
, (2) if %#
and &#
then )(
, and (3) if %#
and &#('
then )#*' for all
2
+' ,! . We write ,when ,#
and ,( z
. We say that
covers and write /.
when 0and there is no ' 1! such that 0-)' and '2.A
finite poset is completely described by the covering relation. The planar Hasse diagram of
"!d
3#q is the graph for which the elements of ! are the nodes and the edges are given by
the covering relation. In addition, we require that if 4.
then is drawn higher than .
The ranking poset
is the poset in which the elements are all possible cosets %65 ,
% . The partial order of
where 7 is an ordered partition of and
is defined by
refinement; that is, &-l if we can get from to by adding vertical lines. Note that
is different from the poset of all set partitions of G
2
ordered by partition refinement
since in
the order of the partition elements matters. Figure 1 shows the Hasse diagram
of
and a portion of the Hasse diagram of .
h
A subposet 8E
$#:9! of "!d
3#<;$ is defined by 8>=?! and &#:9
if and only if @#:; .
A chain is a poset in which every two elements are comparable. A saturated chain A of
.m&&&:. .
length + is a sequence of elements
! that satisfy .
A chain of ! is a maximal chain if there is no other saturated chain of ! that contains it. A
graded poset of rank is a poset in which every maximal chain has length . In a graded
poset, there is a rank or grade function )p!As 3v:
such that (mv if is a
minimal element and !(2 2B if 4. .
It is easy to see that
is a graded poset of rank XJ, and the rank of every element
is the number of vertical lines in its denotation. We use to denote the subposet of
\ . In particular, the elements in the + th grade,
consisting of
all of which are incomparable, are denoted by . Full orderings occupy the topmost
grade . Classifications "
" reside in . Other elements of are
#
multilabel classifications liG
2
where = GG
.
4 Conditional Models on the Ranking Poset
We now present a family of conditional models defined in terms of the ranking poset. To
. That is, o2
( oc
: for
begin, suppose that o is a right invariant function on
all
and
% . Here right invariance is defined with respect to the natural
action of %
on
, given by
[
[
[
T3
TZi I&&&i
[
[
[
&$y(53 T 0
T
0
&&&3i3
I
(4)
The function o may or may not be a metric; its interpretation as a measure of dissimilarity,
however, remains.
We will examine several distances that are based on the covering relation . of
. Down
and up moves on the Hasse diagram will be denoted by and respectively. A distance
o defined in terms of and moves is easily shown to be right invariant because the
group action of %
does not change the covering relation between any two elements; that
is, the group action of %
on
commutes with and moves:
JJJJ0s
JJ JJ0s
JJ:JJ0s
(5)
JJ: JJ0s
While the metric properties of o are not[ required in our model, the right invariance property
is essential since we want to treat all in the same manner.
We are now ready to give the general form of a conditional model on
. Let o be an
invariant function, as above. The model takes as input + rankings
>
contained
in some subset =
of the ranking poset. For example, each ! could be an element of
, which will be the ?carrier density?
# . Let " be a probability mass function on
or default model. Then o and " specify an exponential model $#M given by
c6$#M (
:
9
where & 6587 t ,
normalizing constant
%
'&D
(#d (
%
/0
'&
(#M
)8+*",.-
, and .! ;
=
<.=?>
/0
+*",.-
! 21
! 1
=
!o2
+!$34
.
The term
! oc
!
43
%
(6)
'&
#M is the
(7)
, 7 &3 #M forms a probability distribution over 9(= .
Given a data set c
# , the parameters 5 will[ typically
[ be selected by maxi( 1 [
2 $# , a marginal likelihood
mizing the conditional loglikelihood
&
or posterior. Under mild regularity conditions,
will be convex and have a unique global
1
maximum.
Thus, conditional on #
[
[
4.1 A Bayesian interpretation
We now derive a Bayesian interpretation for the model given by (6). Our result parallels
the interpretation of multistage ranking models given by Fligner and Verducci [6]. The key
fact is that, under appropriate assumptions, the normalizing term does not depend on the
partial ordering in the one-dimensional case.
Proposition 4.1. Suppose that o is right invariant and that
of % . If %
acts transitively on 9 then
for all
7
9 and
1
=
< (
=
is invariant under the action
<
(8)
t .
Proof. First, note that since =
is invariant under the action of % , it follows that
(
% . [Indeed,
7
for each[
by the invariance
assumption,[! and :7
!
[
since for l(< T8 &&&3
we have "!(<3# T0nI&&&i3# 0
such
that "(/ .
acts transitively on 9 , for all
Now, since %
We thus have that
%
1
(
=
(
(
(
Thus, we can write
fact depend on .
%
7
M(
1
%
<
9 there is # .%
such that "#E(
$ .
(9)
<% %
=
<' %
=
<'
=
(by right invariance of & )
(10)
(11)
(by invariance of ( )
(12)
since the normalizing constant for
1
9 does not in
The underlying generative model is given as follows. Assume that
9 is drawn from
the prior "c and that
are independently drawn from generalized Mallows
models
*) c .!
(
%
) < )
(13)
3!
1
where .! . Then under the conditions of Proposition
4.1, we have from Bayes? rule
that the posterior distribution over is given by
- ) ) < )
%
)8,+ ! * ) _ +!
)8
+ ! !30#
- ) 1 . )/ 0< * )
(
(14)
%
<+=?> )8, + ! * ) _ +!
+ ! ! # <+=?> c
1
$#M
(
(15)
We thus have the following characterization of
7&3$#M .
Proposition 4.2. If o is right invariant, is invariant under the action of % , and %
acts
transitively on 9 , then the model
&$#M defined in equation (6) is the posterior under
*)
& , with prior S " .
independent sampling of generalized Mallows models, 2! S
The
[ conditions of this proposition are satisfied, for example, when 91(*%
% :9G% 5 as is assumed in the special cases of the next section.
:9G%
5
and
(
5 Special Cases
This section derives several special cases of model (6), corresponding to existing ensemble methods. The special cases correspond to different choices of 9]
5 and o in the
definition of the model. In each case c is taken to be uniform, though the extension
to non-uniform is immediate. Following [9], the unnormalized versions of all the
models may be easily derived, corresponding to the exponential loss used in boosting.
5.1 Cranking and Mallows
model
Let 5N( t
(9 ( /( (N% , and let oc
be the minimum number of down#
up ( ) moves on the Hasse diagram of needed to bring to . Since adjacent
transpositions of permutations may be identified with a down move followed by an up
move over the Hasse diagram, oc
is equal to Kendall?s Tau ]c
. For example,
]7 ^: b
b2 ^i ( b and the corresponding path in Figure 1 is
^: b
^ b
^i8 b
^iG
b
^: be
^n
be
b ^:e
In this case model (6) becomes the cranking model [10]
6&2 (
- ) T ) < )
%
#V
&2
&
+! '
%
t
(16)
The Bayesian interpretation in this case is well known, and is derived in [6]. The generative
model is independent sampling of ! from a Mallows model whose location parameter
is and whose scale parameter is ! . Other special cases that fall into this category are the
models of Feigin [5] and Critchlow1 and Verducci [3].
5.2 Logistic models
Let 5a( t , 9 (6*(a% 9%
number of up-down
# , and let oc
be the minimum
( ) moves in the Hasse diagram. Since 9 ( ( % :9%
o2
(/o2 %
#
#
:
0%
#!(
v
#
^
if #
!(
otherwise
#2#
(17)
In this case model (6) becomes equivalent to the multiclass generalization of logistic regression. If the normalization constraints in the corresponding convex primal problem are
removed, the model becomes discrete
[ AdaBoost.M2; that is, o2
!8 29Z^ becomes the
(discrete) multiclass weak learner
3v:
in the usual boosting notation. See [9]
for details on the correspondence between exponential models and the unnormalized models that correspond to AdaBoost.
5.3 Error correcting output codes
A more interesting special case of the algebraic structure described in Sections 3 and 4
is where the ensemble method is error correcting output coding (ECOC) [4]. Here we set
#
, l(
9 (,% :9G%
9
, and take the parameter space to be
[
(18)
55(-& Ot ( >q(*&&&G(
and
vG
1
1
1
1
As before, o2
is the minimal number of up-down ( ) moves in the Hasse diagram
needed to bring to .
T
Since ( %
# , the model computes probabilities of classifications . On input , the base rankers output !G 2
?9 , which corresponds to one of the binary
classifiers in ECOC for the appropriate column of the binary coding matrix.
For example,
consider a binary classifier trained on the coding column
v
v:
v:
vG . On an input ,
the classifier outputs 0 or 1, corresponding to the partial rankings ( ^n
7j
k
e
b and
@(*G
b ^n
j:
0kn
, respectively.
Since
@% 9G% #
and
9
o2
(
(
o2 %
^
[ [
[
= i [ =
if #
#
(20)
otherwise.
For example, if (*^:e
b
j:
k
and X(N^n
j:
k
2iG
b , then o2
`(
from the sequence of moves
^:e
b
j:
k
(19)
^: j:
0kn
2iG
b
, as can be seen
^n
j:
k
2iG
b`
(21)
If '(*8 ^
b:
7j:
0kn
and '(
n^
7j
kn
e
b , then o2
( ^ , with the sequence of moves
^n
7j
k
b
^
7j
kn
b ^
7j:
0kn
e b ^
7j:
0kn
e
bM (22)
^n
b
7j
k
[
Since (
! , the exponent of the model becomes ! o2
+! . At test time, the model
1
1 the label corresponding to the partial ranking
1
( arg
, < n6$#M . Now,
thus selects
since is strictly negative, #M is a monotonically decreasing function in ! o2
+! .
1
Equivalence with the ECOC decision rule thus follows from the fact that ! oc
+!nJ+
is the Hamming distance between the appropriate row of the coding matrix and the concatenation of the bits returned from the binary classifiers.
Thus, with the appropriate definitions of 9\
and o , the conditional model on the ranking
poset is a probabilistic formulation of ECOC that yields the same classification decisions.
This suggests ways in which ECOC might be naturally extended. First, relaxing the constraint `(
> (*&&&8( results in a more general model that corresponds to ECOC with a
1 Hamming
1
1
weighted
distance,
or index sensitive ?channel,? where the learned weights may
adapt to the precision of the various base classifiers. Another simple generalization results
from using a nonuniform carrier density .
A further generalization is achieved by considering that
[ [ for a given
[ [ coding[ [ matrix,[ the
[
trained classifier for a given column outputs either = i =
or = =
depending on the input . Allowing the output of the classifier instead to belong to other
grades of
results in a model that corresponds to error correcting output codes with nonbinary codes. While this is somewhat antithetic to the original spirit of ECOC?reducing
multiclass to binary?the base classifiers in ECOC are often multiclass classifiers such as
decision trees in [4]. For such classifiers, the task instead can be viewed as reducing multiclass to partial ranking. Moreover, there need not be an explicit coding matrix. Instead,
the input rankers may output different partial rankings for different inputs, which are then
combined according to model (6). In this way, a different coding matrix is built for each
example in a dynamic manner. Such a scheme may be attractive in bypassing the problem
of designing the coding matrix.
6 Summary
An algebraic framework has been presented for classification and ranking, leading to conditional models on the ranking poset that are defined in terms of an invariant distance or
dissimilarity function. Using the invariance properties of the distances, we derived a generative interpretation of the probabilistic model, which may prove to be useful in model
selection and validation. Through different choices of the components
9]
and o , the
family of models was shown to include as special cases the Mallows model, and the
classifier combination methods used by logistic models, boosting, cranking, and error correcting output codes. In the case of ECOC, the poset framework shows how probabilities
may be assigned to partial rankings in a way that is consistent with the usual definitions of
ECOC , and suggests several natural extensions.
Acknowledgments
We thank D. Critchlow, G. Hulten and J. Verducci for helpful input on the paper. This work
was supported in part by NSF grant CCR-0122581.
References
[1] M. Collins, R. E. Schapire, and Y. Singer. Logistic regression, AdaBoost and Bregman distances. Machine Learning, 48, 2002.
[2] D. E. Critchlow. Metric Methods for Analyzing Partially Ranked Data. Lecture Notes
in Statistics, volume 34, Springer, 1985.
[3] D. E. Critchlow and J. S. Verducci. Detecting a trend in paired rankings. Journal of
the Royal Statistical Society C, 41(1):17?29, 1992.
[4] T. G. Dietterich and G. Bakiri. Solving multiclass learning problems via errorcorrecting codes. Journal of Artificial Intelligence Research, 2:263?286, 1995.
[5] P. D. Feigin. Modeling and analyzing paired ranking data. In M. A. Fligner and
J. S. Verducci, editors, Probability Models and Statistical Analyses for Ranking Data.
Springer, 1992.
[6] M. A. Fligner and J. S. Verducci. Posterior probabilities for a concensus ordering.
Psychometrika, 55:53?63, 1990.
[7] Y. Freund and R. E. Schapire. Experiments with a new boosting algorithm. In International Conference on Machine Learning, 1996.
[8] M. G. Kendall. A new measure of rank correlation. Biometrika, 30, 1938.
[9] G. Lebanon and J. Lafferty. Boosting and maximum likelihood for exponential models. In Advances in Neural Information Processing Systems, 15, 2001.
[10] G. Lebanon and J. Lafferty. Cranking: Combining rankings using conditional probability models on permutations. In International Conference on Machine Learning,
2002.
[11] C. L. Mallows. Non-null ranking models. Biometrika, 44:114?130, 1957.
[12] R. P. Stanley. Enumerative Combinatorics, volume 1. Wadsworth & Brooks/Cole
Mathematics Series, 1986.
| 2146 |@word mild:1 version:2 commute:1 contains:2 score:3 series:1 o2:19 existing:2 assigning:1 john:1 partition:6 generative:3 selected:1 intelligence:1 item:20 lr:2 transposition:3 characterization:1 boosting:5 node:1 location:1 preference:1 detecting:1 cosets:3 psfrag:2 consists:2 prove:1 manner:2 indeed:1 examine:1 grade:4 ecoc:10 decreasing:1 equipped:1 considering:1 becomes:5 begin:2 psychometrika:1 notation:4 underlying:1 moreover:1 mass:1 null:1 interpreted:1 nj:1 thorough:1 every:3 act:3 preferable:1 biometrika:2 classifier:15 grant:1 positive:1 carrier:2 before:1 treat:1 analyzing:2 path:1 might:1 equivalence:1 suggests:3 relaxing:1 unique:1 acknowledgment:1 mallow:8 poset:25 convenient:1 confidence:5 get:1 selection:1 equivalent:3 independently:1 convex:2 identifying:1 correcting:7 m2:1 rule:2 suppose:2 designing:1 pa:2 element:11 trend:1 jk:1 ordering:5 removed:1 topmost:2 multistage:1 dynamic:1 multilabel:1 trained:2 depend:2 reviewing:1 solving:1 algebra:1 learner:1 completely:1 triangle:1 nonbinary:1 easily:2 various:1 derivation:1 separated:1 describe:1 artificial:1 whose:2 say:1 loglikelihood:1 otherwise:3 statistic:1 sequence:4 mb:1 maximal:2 combining:1 regularity:1 produce:1 posets:1 derive:1 depending:1 school:2 borrowed:1 c:2 quotient:2 concensus:1 require:2 fix:1 generalization:4 proposition:4 extension:5 strictly:1 bypassing:1 hold:1 antithetic:1 cb:2 jx:1 label:2 combinatorial:1 sensitive:1 cole:1 weighted:1 hulten:1 derived:3 rank:8 likelihood:2 helpful:1 typically:2 critchlow:5 relation:5 selects:1 arg:1 classification:11 denoted:9 exponent:1 proposes:1 special:12 wadsworth:1 marginal:1 equal:1 sampling:2 represents:1 relabeling:1 consisting:2 replacement:2 saturated:2 primal:1 chain:6 bregman:1 edge:1 partial:24 tree:1 re:1 minimal:2 column:3 modeling:1 cover:1 subset:1 uniform:2 kn:7 combined:3 calibrated:1 density:2 international:2 probabilistic:3 again:1 satisfied:1 guy:1 tz:1 leading:1 coding:9 b2:2 includes:1 matter:1 satisfy:1 combinatorics:1 ranking:49 mv:1 multiplicative:1 view:1 kendall:6 portion:1 bayes:1 parallel:1 ensemble:8 t3:1 correspond:2 yield:1 generalize:1 weak:1 bayesian:5 zy:1 definition:5 naturally:1 proof:1 hamming:3 popular:1 stanley:1 mizing:1 focusing:1 higher:2 verducci:6 adaboost:5 planar:1 specify:1 formulation:1 though:1 correlation:1 logistic:5 building:1 dietterich:1 concept:2 equality:1 assigned:2 hasse:10 symmetric:1 attractive:1 adjacent:4 covering:4 unnormalized:2 generalized:2 bring:3 volume:2 belong:1 interpretation:9 mellon:2 composition:1 discordant:1 mathematics:1 base:7 posterior:4 certain:1 inequality:1 binary:7 wv:1 seen:1 minimum:3 somewhat:1 algebraically:1 ihd:1 monotonically:1 full:6 adapt:1 metasearch:1 paired:2 prediction:3 regression:3 basic:2 cmu:2 metric:3 sometimes:1 normalization:1 achieved:1 addition:5 want:1 diagram:10 ot:1 lafferty:4 spirit:1 integer:1 easy:1 xj:1 associating:1 identified:1 incomparable:1 multiclass:6 ranker:3 whether:1 algebraic:6 returned:1 proceed:1 action:6 useful:1 amount:1 category:1 schapire:2 occupy:1 nsf:1 dotted:1 ccr:1 carnegie:2 discrete:7 write:3 group:3 key:1 pb:1 drawn:3 graph:1 sum:1 cranking:6 family:2 decision:3 investigates:1 comparable:1 bit:1 followed:1 correspondence:1 denotation:1 constraint:2 according:1 combination:2 invariant:11 indexing:1 errorcorrecting:1 taken:2 equation:1 visualization:1 remains:1 needed:3 singer:1 flip:1 generalizes:2 available:2 coset:2 appropriate:5 original:1 denotes:3 top:4 include:1 unifying:2 graded:3 bakiri:1 society:1 move:10 usual:3 distance:15 thank:1 concatenation:1 enumerative:1 code:8 length:2 index:1 negative:1 rise:1 allowing:1 vertical:3 finite:1 immediate:1 extended:1 nonuniform:1 arbitrary:1 qos:1 pair:3 required:1 distinction:1 learned:1 subgroup:1 brook:1 beyond:1 below:1 built:1 including:2 tau:5 royal:1 ranked:9 natural:3 scheme:2 rated:3 ready:1 gf:1 review:1 prior:2 freund:1 loss:1 lecture:1 permutation:10 interesting:1 vg:1 validation:1 conveyed:1 consistent:1 editor:1 row:1 course:1 summary:2 supported:1 side:2 fall:1 default:1 computes:1 reside:1 collection:2 refinement:2 bm:1 lebanon:4 preferred:2 global:1 rul:1 pittsburgh:2 conclude:1 assumed:1 channel:1 permute:1 transitively:3 fligner:3 precision:1 position:5 explicit:1 exponential:4 third:1 down:5 covariate:1 maxi:1 list:1 normalizing:3 derives:2 essential:1 adding:1 dissimilarity:2 easier:1 ordered:6 contained:1 ux:1 partially:3 bo:3 springer:2 corresponds:4 satisfies:2 conditional:16 viewed:1 change:1 reducing:2 called:1 invariance:8 indicating:2 formally:2 collins:1 |
1,259 | 2,147 | Efficient Learning Equilibrium *
Ronen I. Brafman
Computer Science Department
Ben-Gurion University
Beer-Sheva, Israel
email: [email protected]
Moshe Tennenholtz
Computer Science Department
Stanford University
Stanford, CA 94305
e-mail: [email protected]
Abstract
We introduce efficient learning equilibrium (ELE), a normative approach to learning in non cooperative settings. In ELE, the learning algorithms themselves are required to be in equilibrium. In
addition, the learning algorithms arrive at a desired value after
polynomial time, and deviations from a prescribed ELE become irrational after polynomial time. We prove the existence of an ELE
in the perfect monitoring setting, where the desired value is the
expected payoff in a Nash equilibrium. We also show that an ELE
does not always exist in the imperfect monitoring case. Yet, it
exists in the special case of common-interest games. Finally, we
extend our results to general stochastic games.
1
Introduction
Reinforcement learning in the context of multi-agent interaction has attracted the
attention of researchers in cognitive psychology, experimental economics, machine
learning, artificial intelligence, and related fields for quite some time [8, 4]. Much
of this work uses repeated games [3, 5] and stochastic games [10, 9, 7, 1] as models
of such interactions.
The literature on learning in games in game theory [5] is mainly concerned with the
understanding of learning procedures that if adopted by the different agents will
converge at end to an equilibrium of the corresponding game. The game itself may
be known; the idea is to show that simple dynamics lead to rational behavior, as
prescribed by a Nash equilibrium. The learning algorithms themselves are not required to satisfy any rationality requirement; it is what they converge to, if adopted
by all agents that should be in equilibrium. This is quite different from the classical perspective on learning in Artificial Intelligence, where the main motivation
The second author permanent address is: Faculty of Industrial Engineering and Management , Technion, Haifa 32000, Israel. This work was supported in part by Israel Science
Foundation under Grant #91/02-1. The first author is partially supported by the Paul
Ivanier Center for Robotics and Production Management .
for learning stems from the fact that the model of the environment is unknown.
For example, consider a Markov Decision Process (MDP). If the rewards and transition probabilities are known then one can find an optimal policy using dynamic
programming. The major motivation for learning in this context stems from the
fact that the model (i.e. rewards and transition probabilities) is initially unknown.
When facing uncertainty about the game that is played, game-theorists appeal to
a Bayesian approach, which is completely different from a learning approach; the
typical assumption in that approach is that there exists a probability distribution
on the possible games, which is common-knowledge. The notion of equilibrium is
extended to this context of games with incomplete information, and is treated as
the appropriate solution concept. In this context, agents are assumed to be rational
agents adopting the corresponding (Bayes-) Nash equilibrium, and learning is not
an issue.
In this work we present an approach to learning in games, where there is no known
distribution on the possible games that may be played - an approach that appears to
be much more reflective of the setting studied in machine learning and AI and in the
spirit of work on on-line algorithms in computer science. Adopting the framework of
repeated games, we consider a situation where the learning algorithm is a strategy
for an agent in a repeated game. This strategy takes an action at each stage based
on its previous observations, and initially has no information about the identity of
the game being played. Given the above, the following are natural requirements for
the learning algorithms provided to the agents:
1. Individual Rationality: The learning algorithms themselves should be in
equilibrium. It should be irrational for each agent to deviate from its
learning algorithm, as long as the other agents stick to their algorithms,
regardless of the what the actual game is.
2. Efficiency:
(a) A deviation from the learning algorithm by a single agent (while the
other stick to their algorithms) will become irrational (i.e. will lead to
a situation where the deviator 's payoff is not improved) after polynomially many stages.
(b) If all agents stick to their prescribed learning algorithms then the expected payoff obtained by each agent within a polynomial number of
steps will be (close to) the value it could have obtained in a Nash
equilibrium, had the agents known the game from the outset.
A tuple of learning algorithms satisfying the above properties for a given class of
games is said to be an Efficient Learning Equilibrium[ELE]. Notice that the learning
algorithms should satisfy the desired properties for every game in a given class
despite the fact that the actual game played is initially unknown. Such assumptions
are typical to work in machine learning. What we borrow from the game theory
literature is the criterion for rational behavior in multi-agent systems. That is, we
take individual rationality to be associated with the notion of equilibrium. We also
take the equilibrium of the actual (initially unknown) game to be our benchmark
for success; we wish to obtain a corresponding value although we initially do not
know which game is played.
In the remaining sections we formalize the notion of efficient learning equilibrium,
and present it in a self-contained fashion. We also prove the existence of an ELE
(satisfying all of the above desired properties) for a general class of games (repeated
games with perfect monitoring) , and show that it does not exist for another. Our
results on ELE can be generalized to the context of Pareto-ELE (where we wish
to obtain maximal social surplus), and to general stochastic games. These will be
mentioned only very briefly, due to space limitations. The discussion of these and
other issues, as well as proofs of theorems, can be found in the full paper [2].
Technically speaking, the results we prove rely on a novel combination of the socalled folk theorems in economics, and a novel efficient algorithm for the punishment
of deviators (in games which are initially unknown).
2
ELE: Definition
In this section we develop a definition of efficient learning equilibrium. For ease of
exposition, our discussion will center on two-player repeated games in which the
agents have an identical set of actions A. The generalization to n-player repeated
games with different action sets is immediate, but requires a little more notation.
The extension to stochastic games is fully discussed in the full paper [2].
A game is a model of multi-agent interaction. In a game, we have a set of players,
each of whom performs some action from a given set of actions. As a result of the
players' combined choices, some outcome is obtained which is described numerically
in the form of a payoff vector, i.e., a vector of values, one for each of the players.
A common description of a (two-player) game is as a matrix. This is called a game
in strategic form. The rows of the matrix correspond to player 1 's actions and the
columns correspond to player 2's actions. The entry in row i and column j in the
game matrix contains the rewards obtained by the players if player 1 plays his ith
action and player 2 plays his jth action.
In a repeated game (RG) the players playa given game G repeatedly. We can view
a repeated game, with respect to a game G, as consisting of infinite number of
iterations, at each of which the players have to select an action of the game G .
After playing each iteration, the players receive the appropriate payoffs, as dictated
by that game's matrix, and move to a new iteration.
For ease of exposition we normalize both players' payoffs in the game G to be nonnegative reals between and some positive constant Rmax . We denote this interval
(or set) of possible payoffs by P = [0, Rmax].
?
In a perfect monitoring setting, the set of possible histories of length t is (A2 X p2)t,
and the set of possible histories, H, is the union of the sets of possible histories for all
t 2 0, where (A2 x p 2)O is the empty history. Namely, the history at time t consists
of the history of actions that have been carried out so far, and the corresponding
payoffs obtained by the players. Hence, in a perfect monitoring setting, a player
can observe the actions selected and the payoffs obtained in the past, but does not
know the game matrix to start with. In an imperfect monitoring setup, all that a
player can observe following the performance of its action is the payoff it obtained
and the action selected by the other player. The player cannot observe the other
player's payoff. The definition of the possible histories for an agent naturally follows.
Finally, in a strict imperfect monitoring setting, the agent cannot observe the other
agents' payoff or their actions.
Given an RG , a policy for a player is a mapping from H, the set of possible histories ,
to the set of possible probability distributions over A. Hence, a policy determines the
probability of choosing each particular action for each possible history. A learning
algorithm can be viewed as an instance of a policy.
We define the value for player 1 (resp. 2) of a policy profile (1f, p), where 1f is a
policy for player 1 and p is a policy for player 2, using the expected average reward
criterion as follows. Given an RG M and a natural number T, we denote the
expected T -step undiscounted average reward of player 1 (resp. 2) when the players
follow the policy profile (1f,p), by U1 (M,1f,p,T) (resp. U2 (M,1f,p,T)). We define
Ui(M, 1f, p) = liminfT--+oo Ui(M, 1f, p, T) for i = 1,2.
Let M denote a class of repeated games. A policy profile (1f, p) is a learning equilibrium w.r.t. M if'rh' , p',M E M, we have that U1 (M,1f',p) :::; U 1 (M,1f,p), and
U2 (M,1f,p') :::; U2 (M , 1f,p). In this paper we mainly treat the class M of all repeated games with some fixed action profile (i.e. , in which the set of actions available
to all agents is fixed). However, in Section 4 we consider the class of common-interest
repeated games. We shall stick to the assumption that both agents have a fixed ,
identical set A of k actions.
Our first requirement, then, is that learning algorithms will be treated as strategies.
In order to be individually rational they should be the best response for one another.
Our second requirement is that they rapidly obtain a desired value. The definition
of this desired value may be a parameter, the most natural candidate - though not
the only candidate - being the expected payoffs in a Nash equilibrium of the game.
Another appealing alternative will be discussed later.
Formally, let G be a (one-shot) game, let M be the corresponding repeated game,
and let n(G) be a Nash-equilibrium of G. Then, denote the expected payoff of agent
i in n(G) by Nl/i(n(G)).
A policy profile (1f, p) is an efficient learning equilibrium with respect to the class
of games M if for every E > 0, < 8 < 1, there exists some T > 0, where T is
polynomial in ~,~, and k , such that with probability of at least 1 - 8: (1) For
every t 2: T and for every repeated game M E M (and its corresponding one-shot
game, G), Ui(M, 1f , p, t) 2: Nl/i(n(G)) - E for i = 1,2, for some Nash equilibrium
n(G), and (2) If player 1 (resp. 2) deviates from 1f to 1f' (resp. from p to p') in
iteration l, then U1 (M, 1f', p, l + t) :::; U 1 (M, 1f, p, l + t) + E (resp. U2 (M , 1f, p', l + t) :::;
U2 (M, 1f, p, l + t) + E) for every t 2: T.
?
Notice that a deviation is considered irrational if it does not increase the expected
payoff by more than E. This is in the spirit of E-equilibrium in game theory. This
is done mainly for ease of mathematical exposition. One can replace this part of
the definition, while getting similar results, with the requirement of "standard"
equilibrium, where a deviation will not improve the expected payoff, and even with
the notion of strict equilibrium, where a deviation will lead to a decreased payoff.
This will require, however, that we restrict our attention to games where there
exist a Nash equilibrium in which the agents' expected payoffs are higher than their
probabilistic maximin values.
The definition of ELE captures the insight of a normative approach to learning in
non-cooperative settings. We assume that initially the game is unknown, but the
agents will have learning algorithms that will rapidly lead to the values the players
would have obtained in a Nash equilibrium had they known the game. Moreover ,
as mentioned earlier, the learning algorithms themselves should be in equilibrium.
Notice that each agent's behavior should be the best response against the other
agents' behaviors, and deviations should be irrational, regardless of what the actual
(one-shot) game is.
3
Efficient Learning Equilibrium: Existence
Let M be a repeated game in which G is played at each iteration. Let A =
{al' ... , ak} be the set of possible actions for both agents. Finally let there be
an agreed upon ordering over the actions. The basic idea behind the algorithm is
as follows. The agents collaborate in exploring the game. This requires k 2 moves.
Next, each agent computes a Nash equilibrium of the game and follows it. If more
than one equilibrium exists, then the first one according to the natural lexicographic
ordering is used. l If one of the agents does not collaborate in the initial exploration
phase, the other agent "punishes" this agent. We will show that efficient punishment is feasible. Otherwise, the agents have chosen a Nash-equilibrium, and it is
irrational for them to deviate from this equilibrium unilaterally.
This idea combines the so-called folk-theorems in economics [6], and a technique
for learning in zero-sum games introduced in [1]. Folk-theorems in economics deal
with a technique for obtaining some desired behavior by making a threat of employing a punishing strategy against a deviator from that behavior. When both
agents are equipped with corresponding punishing strategies, the desired behavior will be obtained in equilibrium (and the threat will not be materialized - as a
deviation becomes irrational). In our context however , when an agent deviates in
the exploration phase, then the game is not fully known, and hence punishment
is problematic; moreover, we wish the punishment strategy to be an efficient algorithm (both computationally, and in the time a punishment will materialize and
make deviations irrational). These are addressed by having an efficient punishment algorithm that guarantees that the other agent will not obtain more than
its maximin value, after polynomial time, although the game is initially unknown
to the punishing agent. The latter is based on the ideas of our R-max algorithm,
introduced in [1].
More precisely, consider the following algorithm, termed the ELE algorithm.
The ELE algorithm:
Player 1 performs action ai one time after the other for k times, for all i = 1,2, ... , k.
In parallel, player 2 performs the sequence of actions (al' ... ,ak) k times.
If both players behaved according to the above then a Nash equilibrium of the corresponding (revealed) game is computed, and the players behave according
to the corresponding strategies from that point on. If several Nash equilibria exist, one is selected based on a pre-determined lexicographic ordering.
lIn particular, the agents can choose the equilibrium selected by a fixed shared
algorithm.
If one of the players deviated from the above, we shall call this player the adversary
and the other player the agent. Let G be the Rmax-sum game in which the
adversary's payoff is identical to his payoff in the original game, and where
the agent's payoff is Rmax minus the adversary payoffs. Let M denote the
corresponding repeated game. Thus, G is a constant-sum game where the
agent's goal is to minimize the adversary's payoff. Notice that some of
these payoffs will be unknown (because the adversary did not cooperate
in the exploration phase). The agent now plays according to the following
algorithm:
Initialize: Construct the following model M' of the repeated game M, where the
game G is replaced by a game G' where all the entries in the game matrix
are assigned the rewards (R max , 0). 2
In addition, we associate a boolean valued variable with each joint-action
{assumed, known}. This variable is initialized to the value assumed.
Repeat:
Compute and Act: Compute the optimal probabilistic maximin of G'
and execute it.
Observe and update: Following each joint action do as follows : Let a be
the action the agent performed and let a' be the adversary's action.
If (a, a') is performed for the first time, update the reward associated
with (a,a') in G', as observed, and mark it known. Recall- the agent
takes its payoff to be complementary to the (observed) adversary's
payoff.
We can show that the policy profile in which both agents use the ELE algorithm is
indeed an ELE. Thus:
Theorem 1 Let M be a class of repeated games.
w.r.t. M given perfect monitoring.
Then, there exists an ELE
The proof of the above Theorem, contained in the full paper, is non-trivial. It rests
on the ability of the agent to "punish" the adversary quickly, making it irrational
for the adversary to deviate from the ELE algorithm.
4
Imperfect monitoring
In the previous section we discussed the existence of an ELE in the context of the
perfect monitoring setup. This result allows us to show that our concepts provide
not only a normative, but also a constructive approach to learning in general noncooperative environments. An interesting question is whether one can go beyond
that and show the existence of an ELE in the imperfect monitoring case as well.
Unfortunately, when considering the class M of all games, this is not possible.
Theorem 2 There exist classes of games for which an ELE does not exist given
imperfect monitoring.
2The value 0 given to the adversary does not play an important role here.
Proof (sketch): We will consider the class of all 2 x 2 games and we will show
that an ELE does not exist for this class under imperfect monitoring.
Consider the following games:
1. Gl:
6,
o
M= ( 5, -100
2. G2:
M =
0,100 )
1, 500
(6,5,119 0, 1)
1, 10
Notice that the payoffs obtained for a joint action in Gland G 2 are identical for
player 1 and are different for player 2.
The only equilibrium of G 1 is where both players play the second action, leading
to (1,500). The only equilibrium of G2 is where both players play the first action,
leading to (6,9). (These are unique equilibria since they are obtained by removal of
strictly dominated strategies.)
Now, assume that an ELE exists, and look at the corresponding policies of the
players in that equilibrium. Notice that in order to have an ELE, we must visit the
entry (6,9) most of the times if the game is G2 and visit the entry (1 ,500) most of
the times if the game is G 1; otherwise, player 1 (resp. player 2) will not obtain a
high enough value in G2 (resp. Gl), since its other payoffs in G2 (resp. Gl) are
lower than that.
Given the above, it is rational for player 2 to deviate and pretend that the game is
always Gland behave according to what the suggested equilibrium policy tells it
to do in that case. Since the game might be actually G 1, and player 1 cannot tell
the difference, player 2 will be able to lead to playing the second action by both
players for most times also when the game is G2, increasing its payoff from 9 to 10,
contradicting ELE. I
The above result demonstrates that without additional assumptions, one cannot
provide an ELE under imperfect monitoring. However, for certain restricted classes
of games, we can provide an ELE under imperfect monitoring, as we now show.
A game is called a common-interest game if for every joint-action, all agents receive
the same reward. We can show:
Theorem 3 Let M c - i be the class of common-interest repeated games in which the
number of actions each agent has is a. There exists an ELE for M c - i under strict
imperfect monitoring.
Proof (sketch): The agents use the following algorithm: for m rounds , each agent
randomly selects an action. Following this, each agent plays the action that yielded
the best reward. If multiple actions led to the best reward, the one that was used
first is selected. m is selected so that with probability 1 - J every joint-action will
be selected. Using Chernoff bound we can choose m that is polynomial in the size
of the game (which is a k , where k is the number of agents) and in 1/ J.
I
This result improves previous results in this area, such as the combination of Qlearning and fictitious play used in [3]. Not only does it provably converge in
polynomial time, it is also guaranteed, with probability of 1 - J to converge to the
optimal Nash-equilibrium of the game rather than to an arbitrary (and possibly
non-optimal) Nash-equilibrium.
5
Conclusion
We defined the concept of an efficient learning equilibria - a normative criterion for
learning algorithms. We showed that given perfect monitoring a learning algorithm
satisfying ELE exists, while this is not the case under imperfect monitoring. In
the full paper [2] we discuss related solution concepts, such as Pareto ELE. A
Pareto ELE is similar to a (Nash) ELE, except that the requirement of attaining
the expected payoffs of a Nash equilibrium is replaced by that of maximizing social
surplus. We show that there fexists a Pareto-ELE for any perfect monitoring setting,
and that a Pareto ELE does not always exist in an imperfect monitoring setting. In
the full paper we also extend our discussion from repeated games to infinite horizon
stochastic games under the average reward criterion. We show that under perfect
monitoring, there always exists a Pareto ELE in this setting. Please refer to [2] for
additional details and the full proofs.
References
[1] R. I. Brafman and M. Tennenholtz. R-max - a general polynomial time algorithm for near-optimal reinforcement learning. In IJCAI'Ol, 200l.
[2] R. I. Brafman and M. Tennenholtz. Efficient learning equilibrium. Technical
Report 02-06, Dept. of Computer Science, Ben-Gurion University, 2002.
[3] C. Claus and C. Boutilier. The dynamics of reinforcement learning in cooperative multi-agent systems. In Proc. Workshop on Multi-Agent Learning, pages
602- 608, 1997.
[4] I. Erev and A.E. Roth. Predicting how people play games: Reinforcement
learning in games with unique strategy equilibrium. American Economic Review, 88:848- 881, 1998.
[5] D. Fudenberg and D. Levine. The theory of learning in games. MIT Press,
1998.
[6] D. Fudenberg and J. Tirole. Game Theory. MIT Press, 1991.
[7] J. Hu and M.P. Wellman. Multi-agent reinforcement learning: Theoretical
framework and an algorithms. In Proc. 15th ICML , 1998.
[8] L. P. Kaelbling, M. L. Littman, and A. W. Moore. Reinforcement learning: A
survey. Journal of AI Research, 4:237- 285, 1996.
[9] M. L. Littman. Markov games as a framework for multi-agent reinforcement
learning. In Proc. 11th ICML, pages 157- 163, 1994.
[10] L.S. Shapley. Stochastic Games. In Proc. Nat. Acad. Scie. USA, volume 39,
pages 1095- 1100, 1953.
| 2147 |@word briefly:1 faculty:1 polynomial:8 hu:1 minus:1 shot:3 initial:1 contains:1 punishes:1 past:1 yet:1 attracted:1 must:1 gurion:2 update:2 intelligence:2 selected:7 ith:1 scie:1 mathematical:1 become:2 prove:3 consists:1 combine:1 shapley:1 introduce:1 indeed:1 expected:10 behavior:7 themselves:4 multi:7 ol:1 actual:4 little:1 equipped:1 considering:1 increasing:1 becomes:1 provided:1 notation:1 moreover:2 israel:3 what:5 rmax:4 guarantee:1 every:7 act:1 demonstrates:1 stick:4 grant:1 positive:1 engineering:1 treat:1 acad:1 despite:1 ak:2 punish:1 might:1 studied:1 ease:3 unique:2 union:1 procedure:1 area:1 outset:1 pre:1 cannot:4 close:1 context:7 center:2 maximizing:1 roth:1 go:1 attention:2 economics:4 regardless:2 survey:1 insight:1 borrow:1 his:3 unilaterally:1 notion:4 resp:9 rationality:3 play:9 programming:1 us:1 associate:1 satisfying:3 cooperative:3 observed:2 role:1 levine:1 capture:1 ordering:3 mentioned:2 environment:2 nash:17 ui:3 reward:11 littman:2 dynamic:3 irrational:9 technically:1 upon:1 efficiency:1 completely:1 joint:5 artificial:2 tell:2 outcome:1 choosing:1 quite:2 stanford:3 valued:1 otherwise:2 ability:1 itself:1 sequence:1 interaction:3 maximal:1 rapidly:2 description:1 normalize:1 getting:1 ijcai:1 empty:1 requirement:6 undiscounted:1 perfect:9 ben:2 oo:1 develop:1 ac:1 p2:1 c:1 stochastic:6 exploration:3 require:1 generalization:1 extension:1 exploring:1 strictly:1 considered:1 equilibrium:47 mapping:1 major:1 a2:2 proc:4 individually:1 mit:2 lexicographic:2 always:4 rather:1 mainly:3 industrial:1 initially:8 selects:1 provably:1 issue:2 socalled:1 special:1 initialize:1 field:1 construct:1 having:1 chernoff:1 identical:4 look:1 icml:2 report:1 randomly:1 individual:2 replaced:2 phase:3 consisting:1 interest:4 nl:2 wellman:1 behind:1 tuple:1 folk:3 incomplete:1 initialized:1 desired:8 haifa:1 noncooperative:1 theoretical:1 instance:1 column:2 earlier:1 boolean:1 strategic:1 kaelbling:1 deviation:8 entry:4 technion:1 punishment:6 combined:1 probabilistic:2 quickly:1 management:2 choose:2 possibly:1 tirole:1 bgu:1 cognitive:1 american:1 leading:2 attaining:1 satisfy:2 permanent:1 later:1 view:1 performed:2 start:1 bayes:1 parallel:1 minimize:1 il:1 correspond:2 ronen:1 bayesian:1 monitoring:21 researcher:1 history:9 email:1 definition:6 against:2 naturally:1 associated:2 proof:5 rational:5 ele:34 knowledge:1 recall:1 improves:1 formalize:1 agreed:1 surplus:2 actually:1 appears:1 higher:1 follow:1 response:2 improved:1 done:1 though:1 execute:1 stage:2 sketch:2 gland:2 behaved:1 mdp:1 usa:1 concept:4 hence:3 assigned:1 moore:1 deal:1 round:1 game:98 self:1 please:1 criterion:4 generalized:1 performs:3 cooperate:1 novel:2 common:6 volume:1 extend:2 discussed:3 numerically:1 refer:1 theorist:1 ai:3 collaborate:2 had:2 playa:1 showed:1 dictated:1 perspective:1 termed:1 certain:1 success:1 additional:2 converge:4 full:6 multiple:1 stem:2 technical:1 long:1 lin:1 visit:2 basic:1 iteration:5 adopting:2 robotics:2 receive:2 addition:2 addressed:1 interval:1 decreased:1 rest:1 strict:3 claus:1 spirit:2 call:1 reflective:1 near:1 revealed:1 enough:1 concerned:1 psychology:1 restrict:1 imperfect:12 idea:4 economic:1 whether:1 sheva:1 speaking:1 action:37 repeatedly:1 boutilier:1 exist:8 problematic:1 notice:6 materialize:1 shall:2 threat:2 sum:3 uncertainty:1 arrive:1 erev:1 decision:1 bound:1 guaranteed:1 played:6 deviated:1 nonnegative:1 yielded:1 precisely:1 dominated:1 u1:3 prescribed:3 department:2 according:5 materialized:1 combination:2 appealing:1 making:2 restricted:1 computationally:1 discus:1 know:2 end:1 adopted:2 available:1 observe:5 appropriate:2 alternative:1 existence:5 original:1 remaining:1 pretend:1 classical:1 move:2 question:1 moshe:2 strategy:9 said:1 mail:1 whom:1 trivial:1 length:1 setup:2 unfortunately:1 policy:13 unknown:8 observation:1 markov:2 benchmark:1 behave:2 immediate:1 payoff:30 extended:1 situation:2 arbitrary:1 introduced:2 namely:1 required:2 maximin:3 address:1 beyond:1 tennenholtz:3 adversary:10 suggested:1 able:1 max:3 treated:2 natural:4 rely:1 predicting:1 improve:1 carried:1 deviate:6 review:1 literature:2 understanding:1 removal:1 fully:2 interesting:1 limitation:1 fictitious:1 facing:1 foundation:1 agent:56 beer:1 pareto:6 playing:2 production:1 row:2 brafman:4 supported:2 repeat:1 gl:3 jth:1 transition:2 computes:1 author:2 reinforcement:7 far:1 polynomially:1 social:2 employing:1 qlearning:1 assumed:3 ca:1 obtaining:1 punishing:3 did:1 main:1 fudenberg:2 rh:1 motivation:2 paul:1 profile:6 contradicting:1 repeated:19 complementary:1 fashion:1 wish:3 candidate:2 theorem:8 normative:4 appeal:1 exists:9 workshop:1 nat:1 horizon:1 rg:3 led:1 contained:2 partially:1 g2:6 u2:5 determines:1 identity:1 viewed:1 goal:1 exposition:3 replace:1 shared:1 feasible:1 typical:2 infinite:2 determined:1 except:1 called:3 experimental:1 player:47 select:1 formally:1 mark:1 people:1 latter:1 constructive:1 dept:1 |
1,260 | 2,148 | Coulomb Classifiers: Generalizing
Support Vector Machines via an Analogy
to Electrostatic Systems
Sepp Hochreiter? , Michael C. Mozer? , and Klaus Obermayer?
?
Department of Electrical Engineering and Computer Science
Technische Universit?at Berlin, 10587 Berlin, Germany
?
Department of Computer Science
University of Colorado, Boulder, CO 80309?0430, USA
{hochreit,oby}@cs.tu-berlin.de, [email protected]
Abstract
We introduce a family of classifiers based on a physical analogy to
an electrostatic system of charged conductors. The family, called
Coulomb classifiers, includes the two best-known support-vector
machines (SVMs), the ??SVM and the C?SVM. In the electrostatics analogy, a training example corresponds to a charged conductor
at a given location in space, the classification function corresponds
to the electrostatic potential function, and the training objective
function corresponds to the Coulomb energy. The electrostatic
framework provides not only a novel interpretation of existing algorithms and their interrelationships, but it suggests a variety of new
methods for SVMs including kernels that bridge the gap between
polynomial and radial-basis functions, objective functions that do
not require positive-definite kernels, regularization techniques that
allow for the construction of an optimal classifier in Minkowski
space. Based on the framework, we propose novel SVMs and perform simulation studies to show that they are comparable or superior to standard SVMs. The experiments include classification
tasks on data which are represented in terms of their pairwise proximities, where a Coulomb Classifier outperformed standard SVMs.
1
Introduction
Recently, Support Vector Machines (SVMs) [2, 11, 9] have attracted much interest in
the machine-learning community and are considered state of the art for classification
and regression problems. One appealing property of SVMs is that they are based
on a convex optimization problem, which means that a single minimum exists and
can be computed efficiently. In this paper, we present a new derivation of SVMs
by analogy to an electrostatic system of charged conductors. The electrostatic
framework not only provides a physical interpretation of SVMs, but it also gives
insight into some of the seemingly arbitrary aspects of SVMs (e.g., the diagonal of
the quadratic form), and it allows us to derive novel SVM approaches. Although we
are the first to make the analogy between SVMs and electrostatic systems, previous
researchers have used electrostatic nonlinearities in pattern recognition [1] and a
mechanical interpretation of SVMs was introduced in [9].
In this paper, we focus on the classification of an input vector x ? X into one of
two categories, labeled ?+? and ???. We assume a supervised learning paradigm
in which N training examples are available, each example i consisting of an input
xi and a label yi ? {?1, +1}. We will introduce three electrostatic models that
are directly analogous to existing machine-learning (ML) classifiers, each of which
builds on and generalizes the previous. For each model, we describe the physical
system upon which it is based and show its correspondence to an ML classifier.
1.1 Electrostatic model 1: Uncoupled point charges
Consider an electrostatic system of point charges populating a space X 0 homologous
to X . Each point charge corresponds to a particular training example; point charge
i is fixed at location?xi in X 0 , and
? has a charge
? of sign yi . ?We define two sets of
fixed charges: S + = xi | yi = +1 and S ? = xi | yi = ?1 . The charge of point
i is Qi ? yi ?i , where ?i ? 0 is the amount of charge, to be discussed below.
We briefly review some elementary physics. If a unit positive charge is at x in
X 0 , it will be attracted to all charges in S ? and repelled by all charges in S + . To
? the attractive and repelling forces
move the charge from x to some other location x,
must be overcome at every point along the trajectory; the path integral of the force
along the trajectory is called the work and does not depend on the trajectory. The
potential at x is the work that must be done to move a unit positive charge from a
reference point (usually infinity) to x.
? j ?
PN
The potential at x is ? (x) =
j=1 Qj G x , x , where G is a function of the
distance. In electrostatic systems with point charges, G (a, b) = 1/ ka ? bk 2 . From
this definition, one can see that the potential at x is negative (positive) if x is in
a neighborhood of many negative (positive) charges. Thus, the potential indicates
the sign and amount of charge in the local neighborhood.
Turning back to the ML classifier, one might propose a classification rule for some
input x that assigns the label ?+? if ?(x) > 0 or ??? otherwise. Abstracting
from the electrostatic system, if ?i = 1 and G is a function that decreases sufficiently steeply with distance, we obtain a nearest-neighbor classifier. This potential
classifier can be also interpreted as Parzen windows classifier [9].
1.2 Electrostatic model 2: Coupled point charges
Consider now an electrostatic model that extends the previous model in two respects. First, the point charges are replaced by conductors, e.g., metal spheres.
Each conductor i has a self?potential coefficient, denoted Pii , which is a measure
of how much charge it can easily hold; for a metal sphere, Pii is related to sphere?s
diameter. Second, the conductors in S + are coupled, as are the conductors in S ? .
?Coupling? means that charge is free to flow between the conductors. Technically,
S + and S ? can each be viewed as a single conductor.
In this model, we initially place the same charge ?/N on each conductor, and allow
charges within S + and S ? to flow freely (we assume no resistance in the coupling
and no polarization of the conductors). After the charges redistribute, charge will
tend to end up on the periphery of a homogeneous neighborhood of conductors,
because like charges repel. Charge will also tend to end up along the S + ?S ?
boundary because opposite charges attract. Figure 1 depicts the redistribution of
charges, where the shading is proportional to the magnitude ?i . An ML classifier
can be built based on this model, once again using ?(x) > 0 as the decision rule
for classifying an input x. In this model, however, the ?i are not uniform; the
conductors with large ?i will have the greatest influence on the potential function.
Consequently, one can think of ?i as the weight or importance of example i. As we
will show shortly, the examples with ?i > 0 are exactly support vectors of an SVM.
-
+
+
+
+
+
+
+
+
+
+
+
+
-
+ +
+
+
+
+
+
+
+
+
-
-
-
- - - - - - -
Figure 1: Coupled conductor system following charge redistribution. Shading reflects the charge magnitude, and the contour indicates a zero potential.
The redistribution of charges in the electrostatic system is achieved via minimization
of the Coulomb energy. Imagine placing the same total charge magnitude, m, on
S + and S ? by dividing it uniformly among the conductors, i.e., ?i = m/ |S yi |. The
free charge flow in S + and S ? yields a distribution of charges, the ?i , such that
Coulomb energy is minimized.
To introduce Coulomb energy, we begin with some preliminaries. The potential at
conductor i, ?(xi ), which we will denote more compactly as ?i , can be described
PN
in terms of the coefficients of potential Pij [10]: ?i = j=1 Pij Qj , where Pij is the
potential induced on conductor i by charge Qj on conductor j; Pii ? Pij ? 0 and
Pij = Pji . If each conductor i is a metal sphere centered at xi and has radius ri
(radii are enforced to be small enough so that the spheres do not touch? each other),
?
the system can be modeled by a point charge Qi at xi , and Pij = G xi , xj as in
the previous section [10]. The self-potential, Pii , is defined as a function of ri . The
Coulomb energy is defined in terms of the potential on the conductors, ?i :
E =
N
N
1 X
1 T
1X
?i Q i =
Q P Q =
Pij yi yj ?i ?j .
2 i=1
2
2 i,j=1
When the energy minimum is reached, the potential ?i will be the same for all
connected i ? S + (i ? S ? ); we denote this potential ?S + (?S ? ).
Two additional constraints on the system of coupled conductors are necessary in
order to interpret the system in terms of existing machine learning models. First,
the positive and negative potentials must be balanced, i.e., ?S + = ??S ? . This
constraint is achieved by setting the reference point of the potentials ?through
? b,
PN
i
+
?
b = ?0.5 (?S + ?S ), into the potential function: ? (x) = i=1 Qi G x , x + b.
Second, the conductors must be prevented from reversing the sign of their charge,
i.e., ?i ? 0, and from holding more than a quantity C of charge, i.e., ?i ? C. These
requirements can be satisfied in the electrostatic model by disconnecting a conductor
i from the charge flow in S + or S ? when ?i reaches a bound, which will subsequently
freeze its charge. Mathematically, the requirements are satisfied by treating energy
minimization as a constrained optimization problem with 0 ? ?i ? C.
The electrostatic system corresponds to a ??support vector machine (??SVM)
[9]
P
?
with
kernel
G
if
we
set
C
=
1/N
.
The
electrostatic
system
assures
that
+
i =
i?S
P
?
=
0.5
?.
The
identity
holds
because
the
Coulomb
energy
is
exactly
the
?
i
i?S
??SVM quadratic objective function, and the thresholded electrostatic potential
evaluated at a location is exactly the SVM decision rule. The minimization of
potentials differences in the systems S + and S ? corresponds to the minimization
of slack variables in the SVM (slack variables express missing potential due to the
upper bound on ?i ). Mercer?s condition [6], the essence of the nonlinear SVM
theory,Ris equivalent to the fact that continuous electrostatic energy is positive, i.e.,
E =
G (x, z) h (x) h (z) dx dz ? 0. The self-potentials of the electrostatic
system provide an interpretation to the diagonal elements in the quadratic objective
function of the SVM. This interpretation of the diagonal elements allows us to
introduce novel kernels and novel SVM methods, as we discuss later.
1.3 Electrostatic model 3: Coupled point charges with battery
In electrostatic model 2, we control the magnitude of charge applied to S + and S ? .
Although we apply the same charge magnitude to each, we do not have to control
the resulting potentials ?S + and ?S ? , which may be imbalanced. We compensate
for this imbalance via the potential offset b. In electrostatic model 3, we control the
potentials ?S + and ?S + directly by adding a battery to the system. We connect
S + to the positive pole of the battery with potential +1 and S ? to the negative
pole with potential ?1. The battery ensures that ?S + = +1 and ?S ? = ?1 because
charges flow from the battery into or out of the system until the systems take on
the potential of the battery poles. The battery can then be removed. The potential
?i = yi is forced by the battery on conductor i. The total Coulomb energy is the
energy from
P model 2 minus
P the work done by the battery. The work done by the
battery is i?N yi Qi = i?N ?i . The Coulomb energy is
N
N
N
X
X
1 T
1 X
?i =
?i .
Q P Q ?
Pij yi yj ?i ?j ?
2
2 i,j=1
i=1
i=1
This physical system corresponds
to a C?support vector machine (C?SVM) [2, 11].
P
The C?SVM requires that i yi ?i = 0; although this constraint may not be fulfilled
in the system described here, it can be enforced by a slightly different system [4]. A
more straightforward relation to the C?SVM is given in [9] where the authors show
that every ??SVM has the same class boundaries as a C?SVM with appropriate C.
2
Comparison of existing and novel models
2.1 Novel Kernels
The electrostatic perspective makes it easy to understand why SVM algorithms can
break down in high-dimensional spaces: Kernels with rapid fall-off induce small potentials and consequently, almost every conductor retains charge. Because a charged
conductor corresponds to a support vector, the number of support vectors is large,
which leads to two disadvantages: (1) the classification procedure is slow, and (2) the
expected generalization error increases with the number of support vectors [11]. We
therefore should use kernels that do not drop off exponentially. The self?potential
permits the use of kernels that would otherwise be invalid, such as a generalization
?
??l
?
?
?
?
of the electric field: G xi , xj := ?xi ? xj ?2 and G xi , xi := ri?l = Pii , where
ri the radius of the ith sphere. The ri s are increased
to?their maximal values, i.e.
?
until they hit other conductors (ri = 0.5 minj ?xi ? xj ?2 ). These kernels, called
?Coulomb kernels?, are invariant to scaling of the input space in the sense that
scaling does not change the minimum of the objective function. Consequently, such
kernels are appropriate for input data with varying local densities. Figure 2 depicts
a classification task with input regions of varying density. The optimal class boundary is smooth in the low data density regions and has high curvature in regions,
where the data density is high. The classification boundary was constructed using
??
??l/2
?2
?
?
, which is an
a C-SVM with a Plummer kernel G xi , xj := ?xi ? xj ? + ?2
2
approximation to our novel Coulomb kernel but lacks its weak singularities.
Figure 2: Two class data with a dense region and trained with a SVM using the
new kernel. Gray-scales indicate the weights ? support vectors are dark. Boundary
curves are given for the novel kernel (solid), best RBF-kernel SVM which overfits
at high density regions where the resulting boundary goes through a dark circle
(dashed), and optimal boundary (dotted).
2.2 Novel SVM models
Our electrostatic framework can be used to derive novel SVM approaches [4], two
representative examples of which we illustrate here.
2.2.1 ??Support Vector Machine (??SVM):
We can exploit the physical interpretation of Pii as conductor i?s self?potential. The
Pii ?s determine the smoothness of the charge distribution at the energy minimum.
We can introduce a parameter ? to rescale the self potential ? Piinew = ? Piiold .
? controls the complexity of the corresponding SVM. With this modification, and
with C = ?, electrostatic model 3 becomes what we call the ??SVM.
2.2.2 p?Support Vector Machine (p?SVM):
At the Coulomb energy minimum the electrostatic potentials equalize: ? i ? yi =
0, ?i (y is the label vector). This motivates the introduction of potential difference,
2
1
1 T T
1 T
T T
2 kP Q + yk2 = 2 Q P P Q + Q P y + 2 y y as the objective. We obtain
1 T
min
? Y P T P Y ? ? 1T Y P Y ?
?
2
subject to
1T P Y ? = 0 , |?i | ? C,
where 1 is the vector of ones and Y := diag(y). We call this variant of the
optimization problem the potential-SVM (p-SVM). Note that the p-SVM is similar
to the ?empirical kernel map? [9]. However P appears in the objective?s linear term
and the constraints. We classify in a space where P is a dot product matrix. The
constraint 1T P Y ? = 0 ensures that the average potential for each class is equal.
By construction, P T P is positive definite; consequently, this formulation does not
require positive definite kernels. This characteristic is useful for problems in which
the properties of the objects to be classified are described by their pairwise proximities. That is, suppose that instead of representing each input object by an explicit
feature vector, the objects are represented by a matrix which contains a real number indicating the similarity of each object to each other object. We can interpret
the entries of the matrix as being produced by an unknown kernel operating on
unknown feature vectors. In such a matrix, however, positive definiteness cannot
be assured, and the optimal hyperplane must be constructed in Minkowski space.
3
Experiments
UCI Benchmark Repository. For the representative models we have introduced, we perform simulations and make comparisons to standard SVM variants.
All datasets (except ?banana? from [7]) are from the UCI Benchmark Repository
and were preprocessed in [7]. We did 100-fold validation on each data set, restricting
the training set to 200 examples, and using the remainder of examples for testing.
We compared two standard architectures, the C?SVM and the ??SVM, to our novel
architectures: to the ??SVM, to the p?SVM, and to a combination of them, the
??p?SVM. The ??p?SVM is a p?SVM regularized like a ??SVM. We explored the
use of radial basis function (RBF), polynomial (POL), and Plummer (PLU) kernels.
Hyperparameters were determined by 5?fold cross validation on the first 5 training
sets. The search for hyperparameter was not as intensive as in [7].
Table 1 shows the results of our comparisons on the UCI Benchmarks. Our two
novel architectures, the ??SVM and the p?SVM, performed well against the two
existing architectures (note that the differences between the C? and the ??SVM
are due to model selection). As anticipated, the p?SVM requires far fewer support vectors. Additionally, the Plummer kernel appears to be more robust against
hyperparameter and SVM choices than the RBF or polynomial kernels.
C
RBF
POL
PLU
6.4
22.8
6.1
RBF
POL
PLU
33.6
36.0
33.4
RBF
POL
PLU
28.7
33.7
28.8
?
?
p
thyroid
9.4
7.7
5.4
12.6
7.0
13.3
6.2
6.1
5.7
breast?cancer
31.6
33.8 32.4
25.7 29.6 27.1
33.1
33.4 30.6
german
29.3
29.0 27.8
29.6 26.2 31.8
28.5
33.3 27.1
?-p
C
?
8.6
6.9
6.1
21.4
20.4
16.3
19.1
20.4
16.3
33.7
29.1
33.4
13.2
35.3
15.7
36.7
35.0
15.7
?
heart
17.9
19.3
16.3
banana
13.2
11.5
15.7
p
?-p
22.4
23.0
17.4
17.8
19.3
16.3
11.6
22.4
21.9
13.4
11.5
15.7
28.8
26.2
33.3
Table 1: Mean % misclassification on 5 UCI Repository data sets. Each cell in
the table is obtained via 100 replications splitting the data into training and test
sets. The comparison is among five SVMs (the table columns) using three kernel
functions (the table rows). Cells in bold face are the best result for a given data set
and italicized the second and third best.
Pairwise Proximity Data. We applied our p?SVM and the generalized SVM
(G?SVM) [3] to two pairwise-proximity data sets. The first data set, the ?cat cortex? data, is a matrix of connection strengths between 65 cat cortical areas and was
provided by [8], where the available anatomical literature was used to determine
proximity values between cortical areas. These areas belong to four different coarse
brain regions: auditory (A), visual (V), somatosensory (SS), and frontolimbic (FL).
The goal was to classify a given cortical area as belonging to a given region or
not. The second data set, the ?protein? data, is the evolutionary distance of 226 sequences of amino acids of proteins obtained by a structural comparison [5] (provided
by M. Vingron). Most of the proteins are from four classes of globins: hemoglobin-ff
(H-ff), hemoglobin-fi (H-fi), myoglobin (M), and heterogenous globins (GH). The
goal was to classify a protein as belonging to a given globin class or not. As Table 2
shows, our novel architecture, the p?SVM, beats out an existing architecture in the
literature, the G?SVM, on 5 of 8 classification tasks, and ties the G?SVM on 2 of
8; it loses out on only 1 of 8.
Size
G-SVM
G-SVM
G-SVM
p-SVM
p-SVM
p-SVM
Reg.
?
0.05
0.1
0.2
0.6
0.7
0.8
cat
V
18
4.6
4.6
6.1
3.1
3.1
3.1
cortex
A
SS
10
18
3.1 3.1
3.1 6.1
1.5 3.1
1.5 6.1
3.1 4.6
3.1 4.6
FL
19
1.5
1.5
3.1
3.1
1.5
1.5
Reg.
?
0.05
0.1
0.2
300
400
500
protein data
H-? H-?
M
72
72
39
1.3
4.0
0.5
1.8
4.5
0.5
2.2
8.9
0.5
0.4
3.5 0.0
0.4 3.1 0.0
0.4
3.5 0.0
GH
30
0.5
0.9
0.9
0.4
0.9
1.3
Table 2: Mean % misclassifications for the cat-cortex and protein data sets using
the p?SVM and the G?SVM and a range of regularization parameters (indicated in
the column labeled ?Reg.?). The result obtained for the cat-cortex data is via leaveone-out cross validation, and for the protein data is via ten-fold cross validation.
The best result for a given classification problem is printed in bold face.
4
Conclusion
The electrostatic framework and its analogy to SVMs has led to several important
ideas. First, it suggests SVM methods for kernels that are not positive definite.
Second, it suggests novel approaches and kernels that perform as well as standard
methods (will undoubtably perform better on some problems). Third, we demonstrated a new classification technique working in Minkowski space which can be used
for data in form of pairwise proximities. The novel approach treats the proximity
matrix as an SVM Gram matrix which lead to excellent experimental results.
We argued that the electrostatic framework not only characterizes a family of
support-vector machines, but it also characterizes other techniques such as nearest
neighbor classification. Perhaps the most important contribution of the electrostatic framework is that, by interrelating and encompassing a variety of methods,
it lays out a broad space of possible algorithms. At present, the space is sparsely
populated and has barely been explored. But by making the dimensions of this
space explicit, the electrostatic framework allows one to easily explore the space
and discover novel algorithms. In the history of machine learning, such general
frameworks have led to important advances in the field.
Acknowledgments
We thank G. Hinton and J. Schmidhuber for stimulating conversations leading to
this research and an anonymous reviewer who provided helpful advice on the paper.
References
[1] M. A. Aizerman, E. M. Braverman, and L. I. Rozono?er. Theoretical foundations
of the potential function method in pattern recognition learning. Automation
and Remote Control, 25:821?837, 1964.
[2] C. J. C. Burges. A tutorial on support vector machines for pattern recognition.
Data Mining and Knowledge Discovery, 2(2):1?47, 1998.
[3] T. Graepel, R. Herbrich, B. Sch?olkopf, A. J. Smola, P. L. Bartlett, K.-R.
M?
uller, K. Obermayer, and R. C. Williamson. Classification on proximity data
with LP?machines. In Proceedings of the Ninth International Conference on
Artificial Neural Networks, pages 304?309, 1999.
[4] S. Hochreiter and M. C. Mozer. Coulomb classifiers: Reinterpreting SVMs as
electrostatic systems. Technical Report CU-CS-921-01, Department of Computer Science, University of Colorado, Boulder, 2001.
[5] T. Hofmann and J. Buhmann. Pairwise data clustering by deterministic annealing. IEEE Trans. Pattern Anal. and Mach. Intelligence, 19(1):1?14, 1997.
[6] J. Mercer. Functions of positive and negative type and their connection with the
theory of integral equations. Philosophical Transactions of the Royal Society
of London A, 209:415?446, 1909.
[7] G. R?
atsch, T. Onoda, and K.-R. M?
uller. Soft margins for AdaBoost. Technical
Report NC-TR-1998-021, Dep. of Comp. Science, Univ. of London, 1998.
[8] J. W. Scannell, C. Blakemore, and M. P. Young. Analysis of connectivity in
the cat cerebral cortex. The Journal of Neuroscience, 15(2):1463?1483, 1995.
[9] B. Sch?
olkopf and A. J. Smola. Learning with Kernels ? Support Vector Machines, Regularization, Optimization, and Beyond. MIT Press, 2002.
[10] M. Schwartz. Principles of Electrodynamics. Dover Publications, NY, 1987.
Republication of McGraw-Hill Book 1972.
[11] V. Vapnik. The nature of statistical learning theory. Springer, NY, 1995.
| 2148 |@word repository:3 briefly:1 cu:1 polynomial:3 simulation:2 minus:1 solid:1 shading:2 tr:1 contains:1 existing:6 ka:1 repelling:1 dx:1 attracted:2 must:5 hofmann:1 hochreit:1 treating:1 drop:1 intelligence:1 fewer:1 ith:1 dover:1 provides:2 coarse:1 location:4 herbrich:1 five:1 along:3 constructed:2 replication:1 reinterpreting:1 introduce:5 pairwise:6 expected:1 rapid:1 brain:1 window:1 becomes:1 begin:1 provided:3 discover:1 what:1 interpreted:1 every:3 charge:46 tie:1 exactly:3 universit:1 classifier:13 hit:1 schwartz:1 control:5 unit:2 positive:13 engineering:1 local:2 treat:1 mach:1 path:1 might:1 suggests:3 co:1 plu:4 blakemore:1 range:1 acknowledgment:1 yj:2 testing:1 definite:4 procedure:1 area:4 empirical:1 printed:1 radial:2 induce:1 protein:7 cannot:1 selection:1 influence:1 disconnecting:1 equivalent:1 deterministic:1 map:1 charged:4 missing:1 dz:1 sepp:1 straightforward:1 go:1 demonstrated:1 convex:1 reviewer:1 splitting:1 assigns:1 insight:1 rule:3 analogous:1 construction:2 imagine:1 colorado:3 suppose:1 homogeneous:1 element:2 recognition:3 lay:1 sparsely:1 labeled:2 electrical:1 region:7 ensures:2 connected:1 remote:1 decrease:1 removed:1 balanced:1 mozer:3 equalize:1 complexity:1 pol:4 battery:10 trained:1 depend:1 technically:1 upon:1 basis:2 compactly:1 easily:2 represented:2 cat:6 derivation:1 univ:1 forced:1 describe:1 plummer:3 london:2 kp:1 artificial:1 oby:1 klaus:1 neighborhood:3 s:2 otherwise:2 think:1 seemingly:1 sequence:1 propose:2 maximal:1 product:1 remainder:1 tu:1 uci:4 olkopf:2 requirement:2 object:5 derive:2 coupling:2 illustrate:1 rescale:1 nearest:2 dep:1 dividing:1 c:3 indicate:1 pii:7 somatosensory:1 radius:3 subsequently:1 centered:1 redistribution:3 require:2 argued:1 generalization:2 preliminary:1 anonymous:1 elementary:1 singularity:1 mathematically:1 hold:2 proximity:8 sufficiently:1 considered:1 outperformed:1 label:3 bridge:1 reflects:1 minimization:4 uller:2 mit:1 pn:3 varying:2 publication:1 focus:1 indicates:2 steeply:1 sense:1 helpful:1 attract:1 initially:1 relation:1 germany:1 classification:13 among:2 denoted:1 art:1 constrained:1 field:2 once:1 equal:1 placing:1 broad:1 scannell:1 anticipated:1 minimized:1 report:2 hemoglobin:2 replaced:1 consisting:1 undoubtably:1 interest:1 mining:1 braverman:1 redistribute:1 integral:2 necessary:1 circle:1 theoretical:1 increased:1 classify:3 column:2 soft:1 disadvantage:1 retains:1 pole:3 technische:1 entry:1 uniform:1 connect:1 density:5 international:1 physic:1 off:2 michael:1 parzen:1 connectivity:1 again:1 satisfied:2 book:1 leading:1 potential:39 nonlinearities:1 de:1 bold:2 includes:1 coefficient:2 automation:1 later:1 break:1 performed:1 overfits:1 characterizes:2 reached:1 contribution:1 acid:1 characteristic:1 efficiently:1 who:1 yield:1 weak:1 populating:1 produced:1 trajectory:3 comp:1 researcher:1 classified:1 history:1 minj:1 reach:1 definition:1 against:2 energy:14 auditory:1 conversation:1 knowledge:1 graepel:1 back:1 appears:2 supervised:1 adaboost:1 formulation:1 done:3 evaluated:1 smola:2 until:2 working:1 touch:1 nonlinear:1 lack:1 indicated:1 perhaps:1 gray:1 usa:1 regularization:3 polarization:1 attractive:1 self:6 essence:1 generalized:1 hill:1 interrelationship:1 gh:2 novel:17 recently:1 fi:2 superior:1 physical:5 exponentially:1 cerebral:1 myoglobin:1 discussed:1 interpretation:6 belong:1 interpret:2 freeze:1 smoothness:1 populated:1 dot:1 similarity:1 yk2:1 operating:1 cortex:5 electrostatic:34 curvature:1 imbalanced:1 perspective:1 periphery:1 schmidhuber:1 yi:12 minimum:5 additional:1 freely:1 determine:2 paradigm:1 dashed:1 smooth:1 technical:2 cross:3 sphere:6 compensate:1 prevented:1 qi:4 variant:2 regression:1 breast:1 kernel:26 globin:3 hochreiter:2 achieved:2 cell:2 annealing:1 sch:2 induced:1 tend:2 subject:1 flow:5 call:2 structural:1 enough:1 easy:1 variety:2 xj:6 misclassifications:1 architecture:6 opposite:1 idea:1 intensive:1 qj:3 bartlett:1 resistance:1 useful:1 amount:2 dark:2 ten:1 svms:15 category:1 diameter:1 tutorial:1 dotted:1 sign:3 fulfilled:1 neuroscience:1 anatomical:1 hyperparameter:2 express:1 four:2 preprocessed:1 thresholded:1 enforced:2 extends:1 family:3 place:1 almost:1 decision:2 scaling:2 comparable:1 bound:2 fl:2 correspondence:1 fold:3 quadratic:3 strength:1 infinity:1 constraint:5 ri:7 aspect:1 thyroid:1 min:1 minkowski:3 department:3 combination:1 belonging:2 vingron:1 slightly:1 appealing:1 lp:1 modification:1 making:1 invariant:1 boulder:2 heart:1 equation:1 assures:1 slack:2 discus:1 german:1 end:2 available:2 generalizes:1 permit:1 apply:1 coulomb:15 appropriate:2 shortly:1 pji:1 clustering:1 include:1 exploit:1 build:1 society:1 objective:7 move:2 quantity:1 diagonal:3 obermayer:2 evolutionary:1 distance:3 thank:1 berlin:3 italicized:1 barely:1 modeled:1 nc:1 holding:1 repelled:1 negative:5 anal:1 motivates:1 unknown:2 perform:4 upper:1 imbalance:1 datasets:1 benchmark:3 beat:1 hinton:1 banana:2 ninth:1 arbitrary:1 community:1 introduced:2 bk:1 mechanical:1 repel:1 connection:2 philosophical:1 uncoupled:1 heterogenous:1 trans:1 beyond:1 below:1 pattern:4 usually:1 aizerman:1 built:1 including:1 royal:1 greatest:1 misclassification:1 force:2 homologous:1 regularized:1 turning:1 buhmann:1 representing:1 coupled:5 review:1 literature:2 discovery:1 encompassing:1 abstracting:1 proportional:1 analogy:6 validation:4 foundation:1 metal:3 pij:8 mercer:2 principle:1 classifying:1 row:1 cancer:1 free:2 allow:2 understand:1 burges:1 neighbor:2 fall:1 face:2 leaveone:1 overcome:1 boundary:7 curve:1 cortical:3 gram:1 contour:1 dimension:1 author:1 far:1 transaction:1 mcgraw:1 ml:4 xi:15 continuous:1 search:1 why:1 table:7 additionally:1 nature:1 onoda:1 robust:1 williamson:1 excellent:1 electric:1 diag:1 assured:1 did:1 dense:1 hyperparameters:1 amino:1 advice:1 representative:2 ff:2 depicts:2 definiteness:1 slow:1 ny:2 explicit:2 third:2 young:1 down:1 er:1 offset:1 explored:2 svm:59 exists:1 restricting:1 adding:1 vapnik:1 importance:1 magnitude:5 margin:1 gap:1 generalizing:1 led:2 explore:1 visual:1 springer:1 corresponds:8 loses:1 stimulating:1 viewed:1 identity:1 goal:2 consequently:4 invalid:1 rbf:6 change:1 determined:1 except:1 uniformly:1 reversing:1 hyperplane:1 conductor:28 called:3 total:2 experimental:1 atsch:1 indicating:1 support:16 reg:3 |
1,261 | 2,149 | Identity Uncertainty and Citation Matching
Hanna Pasula, Bhaskara Marthi, Brian Milch, Stuart Russell, Ilya Shpitser
Computer Science Division, University Of California
387 Soda Hall, Berkeley, CA 94720-1776
pasula, marthi, milch, russell, [email protected]
Abstract
Identity uncertainty is a pervasive problem in real-world data analysis. It
arises whenever objects are not labeled with unique identifiers or when
those identifiers may not be perceived perfectly. In such cases, two observations may or may not correspond to the same object. In this paper,
we consider the problem in the context of citation matching?the problem of deciding which citations correspond to the same publication. Our
approach is based on the use of a relational probability model to define
a generative model for the domain, including models of author and title
corruption and a probabilistic citation grammar. Identity uncertainty is
handled by extending standard models to incorporate probabilities over
the possible mappings between terms in the language and objects in the
domain. Inference is based on Markov chain Monte Carlo, augmented
with specific methods for generating efficient proposals when the domain
contains many objects. Results on several citation data sets show that
the method outperforms current algorithms for citation matching. The
declarative, relational nature of the model also means that our algorithm
can determine object characteristics such as author names by combining
multiple citations of multiple papers.
1 INTRODUCTION
Citation matching is the problem currently handled by systems such as Citeseer [1]. 1 Such
systems process a large number of scientific publications to extract their citation lists. By
grouping together all co-referring citations (and, if possible, linking to the actual cited
paper), the system constructs a database of ?paper? entities linked by the ?cites(p 1 , p2 )?
relation. This is an example of the general problem of determining the existence of a set
of objects, and their properties and relations, given a collection of ?raw? perceptual data;
this problem is faced by intelligence analysts and intelligent agents as well as by citation
systems.
A key aspect of this problem is determining when two observations describe the same
object; only then can evidence be combined to develop a more complete description of the
object. Objects seldom carry unique identifiers around with them, so identity uncertainty
is ubiquitous. For example, Figure 1 shows two citations that probably refer to the same
paper, despite many superficial differences. Citations appear in many formats and are rife
with errors of all kinds. As a result, Citeseer?which is specifically designed to overcome
such problems?currently lists more than 100 distinct AI textbooks published by Russell
1
See citeseer.nj.nec.com. Citeseer is now known as ResearchIndex.
[Lashkari et al 94] Collaborative Interface Agents, Yezdi Lashkari, Max Metral, and Pattie
Maes, Proceedings of the Twelfth National Conference on Articial Intelligence, MIT Press,
Cambridge, MA, 1994.
Metral M. Lashkari, Y. and P. Maes. Collaborative interface agents. In Conference of the
American Association for Artificial Intelligence, Seattle, WA, August 1994.
Figure 1: Two citations that probably refer to the same paper.
and Norvig on or around 1995, from roughly 1000 citations. Identity uncertainty has been
studied independently in several fields. Record linkage [2] is a method for matching up
the records in two files, as might be required when merging two databases. For each pair
of records, a comparison vector is computed that encodes the ways in which the records
do and do not match up. EM is used to learn a naive-Bayes distribution over this vector
for both matched and unmatched record pairs, so that the pairwise match probability can
then be calculated using Bayes? rule. Linkage decisions are typically made in a greedy
fashion based on closest match and/or a probability threshold, so the overall process is
order-dependent and may be inconsistent. The model does not provide for a principled
way to combine matched records. A richer probability model is developed by Cohen et
al [3], who model the database as a combination of some ?original? records that are correct
and some number of erroneous versions. They give an efficient greedy algorithm for finding
a single locally optimal assignment of records into groups.
Data association [4] is the problem of assigning new observations to existing trajectories
when multiple objects are being tracked; it also arises in robot mapping when deciding if
an observed landmark is the same as one previously mapped. While early data association systems used greedy methods similar to record linkage, recent systems have tried to
find high-probability global solutions [5] or to approximate the true posterior over assignments [6]. The latter method has also been applied to the problem of stereo correspondence,
in which a computer vision system must determine how to match up features observed in
two or more cameras [7]. Data association systems usually have simple observation models (e.g., Gaussian noise) and assume that observations at each time step are all distinct.
More general patterns of identity occur in natural language text, where the problem of
anaphora resolution involves determining whether phrases (especially pronouns) co-refer;
some recent work [8] has used an early form of relational probability model, although with
a somewhat counterintuitive semantics.
Citeseer is the best-known example of work on citation matching [1]. The system groups
citations using a form of greedy agglomerative clustering based on a text similarity metric
(see Section 6). McCallum et al [9] use a similar technique, but also develop clustering
algorithms designed to work well with large numbers of small clusters (see Section 5).
With the exception of [8], all of the preceding systems have used domain-specific algorithms and data structures; the probabilistic approaches are based on a fixed probability
model. In previous work [10], we have suggested a declarative approach to identity uncertainty using a formal language?an extension of relational probability models [11]. Here,
we describe the first substantial application of the approach. Section 2 explains how to
specify a generative probability model of the domain. The key technical point (Section 3)
is that the possible worlds include not only objects and relations but also mappings from
terms in the language to objects in the domain, and the probability model must include a
prior over such mappings. Once the extended model has been defined, Section 4 details the
probability distributions used. A general-purpose inference method is applied to the model.
We have found Markov chain Monte Carlo (MCMC) to be effective for this and other applications (see Section 5); here, we include a method for generating effective proposals based
on ideas from [9]. The system also incorporates an EM algorithm for learning the local
probability models, such as the model of how author names are abbreviated, reordered, and
misspelt in citations. Section 6 evaluates the performance of four datasets originally used
to test the Citeseer algorithms [1]. As well as providing significantly better performance,
our system is able to reason simultaneously about papers, authors, titles, and publication
types, and does a good job of extracting this information from the grouped citations. For
example, an author?s name can be identified more accurately by combining information
from multiple citations of several different papers. The errors made by our system point to
some interesting unmodeled aspects of the citation process.
2 RPMs
Reasoning about identity requires reasoning about objects, which requires at least some of
the expressive power of a first-order logical language. Our approach builds on relational
probability models (RPMs) [11], which let us specify probability models over possible
worlds defined by objects, properties, classes, and relations.
2.1 Basic RPMs
At its most basic, an RPM, as defined by Koller et al [12], consists of
? A set C of classes denoting sets of objects, related by subclass/superclass relations.
? A set I of named instances denoting objects, each an instance of one class.
? A set A of complex attributes denoting functional relations. Each complex attribute A has a domain type Dom[A] ? C and a range type Range[A] ? C.
? A set B of simple attributes denoting functions. Each simple attribute B has a
domain type Dom[B] ? C and a range V al[B].
? A set of conditional probability models P (B|P a[B]) for the simple attributes.
P a[B] is the set of B?s parents, each of which is a nonempty chain of (appropriately typed) attributes ? = A1 . ? ? ? .An .B 0 , where B 0 is a simple attribute. Probability models may be attached to instances or inherited from classes. The parent
links should be such that no cyclic dependencies are formed.
? A set of instance statements, which set the value of a complex attribute to an
instance of the appropriate class.
We also use a slight variant of an additional concept from [11]: number uncertainty, which
allows for multi-valued complex attributes of uncertain cardinality. We define each such
attribute A as a relation rather than a function, and we associate with it a simple attribute #[A] (i.e., the number of values of A) with a domain type Dom[A] and a range
{0, 1, . . . , max #[A]}.
2.2 RPMs for citations
Figure 2 outlines an RPM for the example citations of Figure 1. There are four classes,
the self-explanatory Author, Paper, and Citation, as well as AuthorAsCited, which represents not actual authors, but author names as they appear when cited. Each citation we
wish to match leads to the creation of a Citation instance; instances of the remaining three
classes are then added as needed to fill all the complex attributes. E.g., for the first citation
of Figure 1, we would create a Citation instance C1 , set its text attribute to the string ?Metral M. ...August 1994.?, and set its paper attribute to a newly created Paper
instance, which we will call P1 . We would then introduce max(#[author]) (here only 3,
for simplicity) AuthorAsCited instances (D11 , D12 , and D13 ) to fill the P1 .obsAuthors (i.e.,
observed authors) attribute, and an equal number of Author instances (A 11 , A12 , and A13 )
to fill both the P1 .authors[i] and the D1i .author attributes. (The complex attributes would
be set using instance statements, which would then also constrain the cited authors to be
equal to the authors of the actual paper. 2 ) Assuming (for now) that the value of C1 .parse
2
Thus, uncertainty over whether the authors are ordered correctly can be modeled using probabilistic instance statements.
A11
Author
A12
surname
#(fnames)
fnames
A13
A21
D11
AuthorAsCited
surname
#(fnames)
fnames
author
A22
A23
D12
D13
D21
D22
Paper
D23
Citation
#(authors)
authors
title
publication type
P1
P2
#(obsAuthors)
obsAuthors
obsTitle
parse
C1
C2
text
paper
Figure 2: An RPM for our Citeseer example. The large rectangles represent classes: the
dark arrows indicate the ranges of their complex attributes, and the light arrows lay out
all the probabilistic dependencies of their basic attributes. The small rectangles represent
instances, linked to their classes with thick grey arrows. We omit the instance statements
which set many of the complex attributes.
is observed, we can set the values of all the basic attributes of the Citation and AuthorAsCited instances. (E.g., given the correct parse, D11 .surname would be set to Lashkari,
and D12 .fnames would be set to (Max)). The remaining basic attributes ? those of the
Paper and Author instances ? represent the ?true? attributes of those objects, and their
values are unobserved.
The standard semantics of RPMs includes the unique names assumption, which precludes
identity uncertainty. Under this assumption, any two papers are assumed to be different
unless we know for a fact that they are the same. In other words, although there are many
ways in which the terms of the language can map to the objects in a possible world, only
one of these identity mappings is legal: the one with the fewest co-referring terms. It is then
possible to express the RPM as an equivalent Bayesian network: each of the basic attributes
of each of the objects becomes a node, with the appropriate parents and probability model.
RPM inference usually involves the construction of such a network. The Bayesian network
equivalent to our RPM is shown in Figure 3.
3 IDENTITY UNCERTAINTY
In our application, any two citations may or may not refer to the same paper. Thus, for
citations C1 and C2 , there is uncertainty as to whether the corresponding papers P 1 and P2
are in fact the same object. If they are the same, they will share one set of basic attributes;
A11.
surname
D12.
#(fnames)
D12.
surname
A11.
fnames
D11.
#(fnames)
D12.
fnames
A21.
#(fnames)
A13.
surname
A12.
fnames
A21.
fnames
A13.
fnames
A13.
#(fnames)
D13.
surname
D11.
fnames
D11.
surname
D13.
#(fnames)
C1.
#(authors)
P1.
title
C1.
text
P1.
pubtype
C1.
obsTitle
A21.
surname
A23.
surname
A22.
fnames
D22.
#(fnames)
D12.
surname
D21.
#(fnames)
D22.
fnames
A23.
fnames
A23.
#(fnames)
D23.
surname
D21.
fnames
D13.
fnames
C1.
parse
A22.
#(fnames)
A22.
surname
A12.
#(fnames)
A12.
surname
A11.
#(fnames)
D23.
fnames
D21.
surname
D23.
#(fnames)
C2.
#(authors)
P2.
title
C2.
parse
C2.
text
C2.
obsTitle
P2.
pubtype
Figure 3: The Bayesian network equivalent to our RPM, assuming C 1 6= C2 .
if they are distinct, there will be two sets. Thus, the possible worlds of our probability
model may differ in the number of random variables, and there will be no single equivalent Bayesian network. The approach we have taken to this problem [10] is to extend the
representation of a possible world so that it includes not only the basic attributes of a set
of objects, but also the number of objects n and an identity clustering ?, that is, a mapping
from terms in the language (such as P1 ) to objects in the world. We are interested only
in whether terms co-refer or not, so ? can be represented by a set of equivalence classes of
terms. For example, if P1 and P2 are the only terms, and they co-refer, then ? is {{P1 , P2 }};
if they do not co-refer, then ? is {{P1 }, {P2 }}.
We define a probability model for the space of extended possible worlds by specifying the
prior P (n) and the conditional distribution P (?|n). As in standard RPMs, we assume that
the class of every instance is known. Hence, we Q
can simplify these distributions further
by factoring them by class, so that, e.g., P (?) = C?C P (?C ). We then distinguish two
cases:
? For some classes (such as the citations themselves), the unique names assumptions
remains appropriate. Thus, we define P (?Citation ) to assign a probability of 1.0
to the one assignment where each citation object is unique.
? For classes such as Paper and Author, whose elements are subject to identity uncertainty, we specify P (n) using a high-variance log-normal distribution. 3 Then
we make appropriate uniformity assumptions to construct P (?C ). Specifically, we
assume that each paper is a priori equally likely to be cited, and that each author is
a priori equally likely to write a paper. Here, ?a priori? means prior to obtaining
any information about the object in question, so the uniformity assumption is entirely reasonable. With these assumptions, the probability of an assignment ? C,k,m
that maps k named instances to m distinct objects, when C contains n objects, is
given by
1
n!
P (?C,k,m ) =
(n ? m)! nk
When n > m, the world contains objects unreferenced by any of the terms. However, these filler objects are obviously irrelevant (if they affected the attributes of
some named term, they would have been named as functions of that term.) Therefore, we never have to create them, or worry about their attribute values.
Our model assumes that the cardinalities and identity clusterings of the classes are independent of each other, as well as of the attribute values. We could remove these assumptions.
For one, it would be straightforward to specify a class-wise dependency model for n or ?
using standard Bayesian network semantics, where the network nodes correspond to the
cardinality attributes of the classes. E.g., it would be reasonable to let the total number of
papers depend on the total number of authors. Similarly, we could allow ? to depend on the
attribute values?e.g., the frequency of citations to a given paper might depend on the fame
of the authors?provided we did not introduce cyclic dependencies.
4 The Probability Model
We will now fill in the details of the conditional probability models. Our priors over the
?true? attributes are constructed off-line, using the following resources: the 1990 Census data on US names, a large A.I. BibTeX bibliography, and a hand-parsed collection of
500 citations. We learn several bigram models (actually, linear combinations of a bigram
model and a unigram model): letter-based models of first names, surnames, and title words,
as well as higher-level models of various parts of the citation string. More specifically, the
values of Author.fnames and Author.surname are modeled as having a a 0.9 chance of being
3
Other models are possible; for example, in situations where objects appear and disappear, P (?)
can be modeled implicitly by specifying the arrival, transition, and departure rates [6].
drawn from the relevant US census file, and a 0.1 chance of being generated using a bigram
model learned from that file. The prior over Paper.titles is defined using a two-tier bigram
model constructed using the bibliography, while the distributions over Author.#(fnames),
Paper.#(authors), and Paper.pubType 4 are derived from our hand-parsed file. The conditional distributions of the ?observed? variables given their true values (i.e., the corruption models of Citation.obsTitle, AuthorAsCited.surname, and AuthorAsCited.fnames) are
modeled as noisy channels where each letter, or word, has a small probability of being
deleted, or, alternatively, changed, and there is also a small probability of insertion. AuthorAsCited.fnames may also be abbreviated as an initial. The parameters of the corruption
models are learnt online, using stochastic EM.
Let us now return to Citation.parse, which cannot be an observed variable, since citation
parsing, or even citation subfield extraction, is an unsolved problem. It is therefore fortunate that our approach lets us handle uncertainty over parses so naturally. The state space
of Citation.parse has two different components. First of all, it keeps track of the citation
style, defined as the ordering of the author and title subfields, as well as the format in which
the author names are written. The prior over styles is learned using our hand-segmented
file. Secondly, it keeps track of the segmentation of Citation.text, which is divided into
an author segment, a title segment, and three filler segments (one before, one after, and
one in between.) We assume a uniform distribution over segmentations. Citation.parse
greatly constrains Citation.text: the title segment of Citation.text must match the value of
Citation.obsTitle, while its author segment must match the combined values of the simple
attributes of Citation.obsAuthors. The distributions over the remaining three segments of
Citation.text are defined using bigram models, with the model used for the final segment
chosen depending on the publication type. These models were, once more, learned using
our pre-segmented file.
5 INFERENCE
With the introduction of identity uncertainty, our model grows from a single Bayesian
network to a collection of networks, one for each possible value of ?. This collection can be
rather large, since the number of ways in which a set can be partitioned grows very quickly
with the size of the set. 5 Exact inference is, therefore, impractical. We use an approximate
method based on Markov chain Monte Carlo.
5.1 MARKOV CHAIN MONTE CARLO
MCMC [13] is a well-known method for approximating an expectation over some distribution ?(x), commonly used when the state space of x is too large to sum over. The weighted
sum over the values of x is replaced by a sum over samples from ?(x), which are generated
using a Markov chain constructed to have ?(x) as a stationary distribution.
There are several ways of building up an appropriate Markov chain. In the Metropolis?
Hastings method (M-H), transitions in the chain are constructed in two steps. First, a
candidate next state x0 is generated from the current state x, using the (more or less arbitrary) proposal distribution q(x0 |x). The probability that
to x0 is actually made is
the move
0
)q(x|x0 )
the acceptance probability, defined as ?(x0 |x) = min 1, ?(x
?(x)q(x0 |x) .
Such a Markov chain will have the right stationary distribution ?(x) as long as q is defined
in such a way that the chain is ergodic. It is even possible to factor q into separate proposals
for various subsets of variables. In those situations, the variables that are not changed by the
transition cancel in the ratio ?(x0 )/?(x), so the required calculation can be quite simple.
4
Publication types range over {article, conference paper, book, thesis, and tech report}
This sequence is described by the Bell numbers, whose asymptotic behaviour is more than exponential.
5
5.2 THE CITATION-MATCHING ALGORITHM
The state space of our MCMC algorithm is the space of all the possible worlds, where
each possible world contains an identity clustering ?, a set of class cardinalities n, and the
values of all the basic attributes of all the objects. Since the ? is given in each world, the
distribution over the attributes can be represented using a Bayesian network as described
in Section 3. Therefore, the probability of a state is simply the product pf P (n), P (?), and
the probability of the hidden attributes of the network.
Our algorithm uses a factored q function. One of our proposals attempts to change n using
a simple random walk. The other suggests, first, a change to ?, and then, values for all the
hidden attributes of all the objects (or clusters in ?) affected by that change. The algorithm
for proposing a change in ?C works as follows:
Select two clusters a1 , a2 ? ?C 6
Create two empty clusters b1 and b2
place each instance i ? a1 ? a2 u.a.r. into b1 or b2
Propose ?0C = ?C ? {a1, a2} ? {b1, b2}
Given a proposed ?0C , suggesting values for the hidden attributes boils down to recovering
their true values from (possibly) corrupt observations, e.g., guessing the true surname of
the author currently known both as ?Simth? and ?Smith?. Since our title and name noise
models are symmetric, our basic strategy is to apply these noise models to one of the
observed values. In the case of surnames, we have the additional resource of a dictionary
of common names, so, some of the time, we instead pick one of the set of dictionary entries
that are within a few corruptions of our observed names. (One must, of course, careful
to account for this hybrid approach in our acceptance probability calculations.) Parses are
handled differently: we preprocess each citation, organizing its plausible segmentations
into a list ordered in terms of descending probability. At runtime, we simply sample from
these discrete distributions. Since we assume that boundaries occur only at punctuation
marks, and discard segmentations of probability < 10?6 , the lists are usually quite short. 7
The publication type variables, meanwhile, are not sampled at all. Since their range is so
small, we sum them out.
5.3 SCALING UP
One of the acknowledged flaws of the MCMC algorithm is that it often fails to scale. In
this application, as the number of papers increases, the simplest approach ? one where
the two clusters a1 and a2 are picked u.a.r ? is likely to lead to many rejected proposals,
as most pairs of clusters will have little in common. The resulting Markov chain will mix
slowly. Clearly, we would prefer to focus our proposals on those pairs of clusters which are
actually likely to exchange their instances. We have implemented an approach based on the
efficient clustering algorithm of McCallum et al [9], where a cheap distance metric is used
to preprocess a large dataset and fragment it into many canopies, or smaller, overlapping
sets of elements that have a non-zero probability of matching. We do the same, using
word-matching as our metric, and setting the thresholds to 0.5 and 0.2. Then, at runtime,
our q(x0 |x) function proposes first a canopy c, and then a pair of clusters u.a.r. from c.
(q(x|x0 ) is calculated by summing over all the canopies which contain any of the elements
of the two clusters.)
6 EXPERIMENTAL RESULTS
We have applied the MCMC-based algorithm to the hand-matched datasets used in [1].
(Each of these datasets contains several hundred citations of machine learning papers, about
half of them in clusters ranging in size from two to twenty-one citations.) We have also
6
7
Note that if the same cluster is picked twice, it will probably be split.
It would also be possible to sample directly from a model such as a hierarchical HMM
Face
Reinforcement
Reasoning
Constraint
349 citations, 242 papers
406 citations, 148 papers
514 citations, 296 papers 295 citations, 199 papers
Phrase matching
94%
79%
86%
89%
RPM + MCMC
97%
94%
96%
93%
Table 1: Results on four Citeseer data sets, for the text matching and MCMC algorithms.
The metric used is the percentage of actual citation clusters recovered perfectly; for the
MCMC-based algorithm, this is an average over all the MCMC-generated samples.
implemented their phrase matching algorithm, a greedy agglomerative clustering method
based on a metric that measures the degrees to which the words and phrases of any two
citations overlap. (They obtain their ?phrases? by segmenting each citation at all punctuation marks, and then taking all the bigrams of all the segments longer than two words.)
The results of our comparison are displayed in Figure 1, in terms of the Citeseer error metric. Clearly, the algorithm we have developed easily beats our implementation of phrase
matching.
We have also applied our algorithm to a large set of citations referring to the textbook Artificial Intelligence: A Modern Approach. It clusters most of them correctly, but there are a
couple of notable exceptions. Whenever several citations share the same set of unlikely errors, they are placed together in a separate cluster. This occurs because we do not currently
model the fact that erroneous citations are often copied from reference list to reference
list, which could be handled by extending the model to include a copiedFrom attribute.
Another possible extension would be the addition of a topic attribute to both papers and authors: tracking the authors? research topics might enable the system to distinguish between
similarly-named authors working in different fields. Generally speaking, we expect that
relational probabilistic languages with identity uncertainty will be a useful tool for creating
knowledge from raw data.
References
[1]
[2]
[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
S. Lawrence, K. Bollacker, and C. Lee Giles. Autonomous citation matching. In Agents, 1999.
I. Fellegi and A. Sunter. A theory for record linkage. In JASA, 1969.
W. Cohen, H. Kautz, and D. McAllester. Hardening soft information sources. In KDD, 2000.
Y. Bar-Shalom and T. E. Fortman. Tracking and Data Association. Academic Press, 1988.
I. J. Cox and S. Hingorani. An efficient implementation and evaluation of Reid?s multiple
hypothesis tracking algorithm for visual tracking. In IAPR-94, 1994.
H. Pasula, S. Russell, M. Ostland, and Y. Ritov. Tracking many objects with many sensors. In
IJCAI-99, 1999.
F. Dellaert, S. Seitz, C. Thorpe, and S. Thrun. Feature correspondence: A markov chain monte
carlo approach. In NIPS-00, 2000.
E. Charniak and R. P. Goldman. A Bayesian model of plan recognition. AAAI, 1993.
A. McCallum, K. Nigam, and L. H. Ungar. Efficient clustering of high-dimensional data sets
with application to reference matching. In KDD-00, 2000.
H. Pasula and S. Russell. Approximate inference for first-order probabilistic languages. In
IJCAI-01, 2001.
A. Pfeffer. Probabilistic Reasoning for Complex Systems. PhD thesis, Stanford, 2000.
A. Pfeffer and D. Koller. Semantics and inference for recursive probability models. In
AAAI/IAAI, 2000.
W.R. Gilks, S. Richardson, and D.J. Spiegelhalter. Markov chain Monte Carlo in practice.
Chapman and Hall, London, 1996.
| 2149 |@word cox:1 version:1 bigram:6 twelfth:1 grey:1 seitz:1 tried:1 citeseer:9 pick:1 maes:2 carry:1 initial:1 cyclic:2 contains:5 fragment:1 charniak:1 denoting:4 bibtex:1 outperforms:1 existing:1 current:2 com:1 recovered:1 assigning:1 must:5 parsing:1 written:1 kdd:2 cheap:1 remove:1 designed:2 stationary:2 generative:2 intelligence:4 greedy:5 half:1 mccallum:3 smith:1 short:1 record:10 node:2 c2:7 constructed:4 consists:1 rife:1 combine:1 ostland:1 introduce:2 x0:8 pairwise:1 roughly:1 p1:10 themselves:1 multi:1 goldman:1 actual:4 little:1 pf:1 cardinality:4 becomes:1 provided:1 matched:3 kind:1 string:2 textbook:2 developed:2 proposing:1 finding:1 unobserved:1 nj:1 impractical:1 berkeley:2 every:1 subclass:1 runtime:2 omit:1 appear:3 reid:1 segmenting:1 before:1 local:1 despite:1 might:3 twice:1 studied:1 equivalence:1 specifying:2 suggests:1 co:6 range:7 subfields:1 d13:5 unique:5 camera:1 gilks:1 recursive:1 practice:1 bell:1 significantly:1 matching:15 word:6 pre:1 cannot:1 milch:2 context:1 descending:1 equivalent:4 map:2 straightforward:1 independently:1 d12:7 resolution:1 d22:3 simplicity:1 ergodic:1 factored:1 rule:1 counterintuitive:1 fill:4 handle:1 autonomous:1 a13:5 norvig:1 construction:1 exact:1 us:1 hypothesis:1 associate:1 element:3 recognition:1 lay:1 labeled:1 database:3 observed:8 pfeffer:2 ordering:1 russell:5 lashkari:4 principled:1 substantial:1 insertion:1 constrains:1 dom:3 uniformity:2 depend:3 segment:8 iapr:1 creation:1 reordered:1 division:1 easily:1 differently:1 represented:2 various:2 fewest:1 distinct:4 describe:2 effective:2 monte:6 london:1 artificial:2 whose:2 richer:1 quite:2 valued:1 plausible:1 stanford:1 precludes:1 grammar:1 richardson:1 noisy:1 final:1 online:1 obviously:1 sequence:1 ilyas:1 propose:1 product:1 d21:4 relevant:1 combining:2 pronoun:1 organizing:1 description:1 seattle:1 parent:3 cluster:14 empty:1 extending:2 ijcai:2 generating:2 a11:4 object:33 depending:1 develop:2 a22:4 job:1 p2:8 recovering:1 c:1 involves:2 indicate:1 implemented:2 differ:1 thick:1 correct:2 attribute:40 stochastic:1 a12:5 enable:1 mcallester:1 unreferenced:1 explains:1 exchange:1 surname:20 assign:1 behaviour:1 ungar:1 brian:1 secondly:1 extension:2 fellegi:1 around:2 hall:2 normal:1 deciding:2 lawrence:1 mapping:6 dictionary:2 early:2 a2:4 purpose:1 perceived:1 currently:4 title:11 grouped:1 create:3 tool:1 weighted:1 mit:1 clearly:2 sensor:1 gaussian:1 rather:2 publication:7 pervasive:1 derived:1 focus:1 greatly:1 tech:1 inference:7 flaw:1 dependent:1 factoring:1 typically:1 unlikely:1 explanatory:1 hidden:3 relation:7 koller:2 interested:1 semantics:4 overall:1 priori:3 proposes:1 plan:1 field:2 construct:2 once:2 equal:2 never:1 having:1 articial:1 anaphora:1 represents:1 stuart:1 extraction:1 cancel:1 chapman:1 report:1 intelligent:1 simplify:1 few:1 thorpe:1 modern:1 simultaneously:1 national:1 replaced:1 attempt:1 acceptance:2 evaluation:1 punctuation:2 d11:6 light:1 chain:13 d1i:1 unless:1 walk:1 uncertain:1 instance:21 soft:1 giles:1 assignment:4 phrase:6 subset:1 entry:1 uniform:1 hundred:1 too:1 dependency:4 learnt:1 canopy:3 combined:2 referring:3 cited:4 probabilistic:7 off:1 lee:1 together:2 ilya:1 quickly:1 thesis:2 aaai:2 possibly:1 slowly:1 unmatched:1 book:1 american:1 creating:1 shpitser:1 return:1 style:2 suggesting:1 account:1 b2:3 includes:2 notable:1 picked:2 linked:2 bayes:2 inherited:1 kautz:1 collaborative:2 formed:1 variance:1 characteristic:1 who:1 correspond:3 preprocess:2 raw:2 bayesian:8 accurately:1 carlo:6 trajectory:1 corruption:4 published:1 whenever:2 evaluates:1 typed:1 frequency:1 naturally:1 unsolved:1 boil:1 sampled:1 newly:1 dataset:1 couple:1 iaai:1 logical:1 knowledge:1 ubiquitous:1 segmentation:4 actually:3 worry:1 originally:1 higher:1 specify:4 ritov:1 rejected:1 pasula:4 hand:4 hastings:1 working:1 parse:8 expressive:1 overlapping:1 sunter:1 scientific:1 grows:2 name:12 building:1 concept:1 true:6 contain:1 hence:1 symmetric:1 self:1 outline:1 complete:1 interface:2 reasoning:4 ranging:1 wise:1 common:2 functional:1 tracked:1 cohen:2 attached:1 linking:1 association:5 slight:1 extend:1 refer:7 cambridge:1 ai:1 seldom:1 similarly:2 language:9 robot:1 similarity:1 longer:1 closest:1 posterior:1 recent:2 irrelevant:1 shalom:1 discard:1 additional:2 somewhat:1 preceding:1 determine:2 multiple:5 mix:1 segmented:2 technical:1 match:7 academic:1 calculation:2 long:1 divided:1 equally:2 a1:5 variant:1 basic:10 vision:1 metric:6 expectation:1 represent:3 c1:8 proposal:7 addition:1 source:1 appropriately:1 probably:3 file:6 subject:1 inconsistent:1 incorporates:1 call:1 extracting:1 split:1 perfectly:2 identified:1 idea:1 whether:4 handled:4 linkage:4 stereo:1 speaking:1 dellaert:1 generally:1 useful:1 dark:1 locally:1 simplest:1 percentage:1 correctly:2 track:2 write:1 discrete:1 affected:2 express:1 group:2 key:2 four:3 threshold:2 acknowledged:1 drawn:1 deleted:1 a23:4 rectangle:2 sum:4 letter:2 uncertainty:15 soda:1 named:5 place:1 reasonable:2 decision:1 prefer:1 rpm:14 scaling:1 entirely:1 distinguish:2 correspondence:2 copied:1 occur:2 constraint:1 constrain:1 bibliography:2 encodes:1 aspect:2 min:1 format:2 combination:2 smaller:1 em:3 partitioned:1 metropolis:1 census:2 taken:1 tier:1 legal:1 resource:2 previously:1 remains:1 abbreviated:2 nonempty:1 needed:1 know:1 apply:1 hierarchical:1 appropriate:5 hingorani:1 existence:1 original:1 assumes:1 clustering:8 include:4 remaining:3 parsed:2 especially:1 build:1 disappear:1 approximating:1 move:1 added:1 question:1 occurs:1 strategy:1 guessing:1 distance:1 link:1 mapped:1 separate:2 entity:1 landmark:1 hmm:1 thrun:1 topic:2 agglomerative:2 declarative:2 reason:1 analyst:1 assuming:2 modeled:4 providing:1 ratio:1 statement:4 implementation:2 twenty:1 observation:6 markov:10 datasets:3 displayed:1 beat:1 situation:2 relational:6 extended:2 arbitrary:1 august:2 unmodeled:1 pair:5 required:2 california:1 marthi:2 learned:3 nip:1 able:1 suggested:1 bar:1 usually:3 pattern:1 departure:1 including:1 max:4 power:1 overlap:1 natural:1 hybrid:1 spiegelhalter:1 created:1 extract:1 naive:1 faced:1 text:11 prior:6 determining:3 asymptotic:1 subfield:1 par:2 expect:1 interesting:1 agent:4 degree:1 jasa:1 article:1 corrupt:1 share:2 course:1 changed:2 placed:1 formal:1 allow:1 face:1 taking:1 overcome:1 calculated:2 boundary:1 world:12 transition:3 author:39 collection:4 made:3 commonly:1 reinforcement:1 citation:66 approximate:3 implicitly:1 keep:2 global:1 d23:4 b1:3 summing:1 assumed:1 alternatively:1 table:1 channel:1 nature:1 learn:2 ca:1 superficial:1 obtaining:1 nigam:1 hanna:1 complex:9 meanwhile:1 domain:9 did:1 arrow:3 noise:3 arrival:1 identifier:3 augmented:1 fashion:1 fails:1 wish:1 a21:4 exponential:1 fortunate:1 candidate:1 perceptual:1 bhaskara:1 down:1 erroneous:2 specific:2 unigram:1 list:6 evidence:1 grouping:1 merging:1 phd:1 nec:1 fame:1 nk:1 simply:2 likely:4 visual:1 ordered:2 tracking:5 bollacker:1 cite:1 chance:2 ma:1 conditional:4 superclass:1 identity:17 careful:1 change:4 specifically:3 total:2 experimental:1 exception:2 select:1 mark:2 latter:1 arises:2 filler:2 incorporate:1 mcmc:9 researchindex:1 |
1,262 | 215 | Pulse-Firing Neural Chips for Hundreds of Neurons
PULSE-FIRING NEURAL CIDPS
FOR HUNDREDS OF NEURONS
Michael Brownlow
Lionel Tarassenko
Dept. Eng. Science
Univ. of Oxford
Oxford OX1 3PJ
Alan F. Murray
Dept. Electrical Eng.
Univ. of Edinburgh
Mayfield Road
Edinburgh EH9 3JL
Alister Hamilton
II Song Han(l)
H. Martin Reekie
Dept. Electrical Eng.
U niv. of Edinburgh
ABSTRACT
We announce new CMOS synapse circuits using only three
and four MOSFETsisynapse. Neural states are asynchronous
pulse streams, upon which arithmetic is performed directly.
Chips implementing over 100 fully programmable synapses
are described and projections to networks of hundreds of
neurons are made.
1 OVERVIEW OF PULSE FIRING NEURAL VLSI
The inspiration for the use of pulse firing in silicon neural networks is
clearly the electrical/chemical pulse mechanism in "real" biological neurons.
Asynchronous, digital voltage pulses are used to signal states t Si ) through
synapse weights { Tij } to emulate neural dynamics. Neurons fire voltage
pulses of a frequency determined by their level of activity but of a constant
magnitude (usually 5 Volts) [Murray,1989a]. As indicated in Fig. 1,
synapses perform arithmetic directly on these asynchronous pulses, to
increment or decrement the receiving neuron's activity. The activity of a
receiving neuron i, Xi is altered at a frequency controlled by the sending
neuron j, with state Sj by an amount determined by the synapse weight
(here, T ij ).
1 On secondment from the Korean Telecommunications Authority
785
786
Brownlow, Tarassenko, Murray, Hamilton, Han and Reekie
Sj> 0
lij > 0
Sj> 0
lij < 0
Sj = 0
lij > 0
Sj > 0
Tij = 0
Sj > 0
lij < 0
I
S.I
x?
t=:=
I
.fL.fL.n-IL
veo
Figure 1 : Pulse stream synapse functionality
A silicon neural network based on this technique is therefore an
asynchronous, analog computational structure. It is a hybrid between
analog and digital techniques in that the individual neural pulses are digital
voltage spikes, with all the robustness to noise and ease of regeneration that
this implies. These and other characteristics of pulse stream networks will
be discussed in detail later in this paper. Pulse stream methods, developed
in Edinburgh, have since been investigated by other groups - see for
instance [EI-Leithy,1988, Daniell, 1989].
1.1. WHY PULSE STREAMS?
There are some advantages in the use of pulse streams, and pulse rate
encoding, in implementing neural networks. It should be admitted here that
the initial move towards pulse streams was motivated by the desire to
implement pseudo-analog circuits on an essentially digital CMOS process. It
was a decision based at the time on expediency rather than on great vision
on our part, as we did not initially appreciate the full benefits of this form
of pulse stream arithmetic [Murray,1987].
Pulse-Firing Neural Chips for Hundreds of Neurons
For example, the voltages on the terminals of a MOSFET, VGS and VDS
could clearly be used to code a neural synapse weight and state respectively,
doing away with the need for pulses. In the pulse stream form, however,
we can arrange that only VGS is an "unknown". The device equations are
therefore easily simplified, and furthermore the body effect is more
predictable. In an equivalent continuous - time circuit, VDS will also be a
variable, which codes information. Predicting the transistor's operating
regime becomes more difficult, and the equation cannot be simplified.
Aside of the transistor - level advantages, giving rise to extremely compact
synapse circuits, there may be architectural advantages. There are certainly
architectural consequences. Digital pulses are easier to regenerate, easier to
pass between chips, and generally far more noise - insensitive than analog
voltages, all of which are significant advantages in the VLSI context.
Furthermore, the relationship to the biological exemplar should not be
ignored. It is at least interesting - whether it is significant remains to be
seen.
2 FULLY ANALOG PULSE STREAM SYNAPSES
Our early pulse stream chips proved the viability of the pulse stream
technique [Murray,1988a]. However, the area occupied by the digital
weight storage memory was unacceptably large. Furthermore, the use of
pseudo-clocks in an analog circuit was both aesthetically unsatisfactory and
detrimental to smooth dynamical behaviour, and using separate signal paths
for excitation and inhibition was both clumsy and inefficient. Accordingly,
we have developed a family of fully programmable, fully analog synapses
using dynamic weight storage, and operating on individual pulses to perform
arithmetic. We have already reported time-modulation synapses based on
this technique, and a later section of this paper will present the associated
chips [Murray,1988b, Murray,1989b].
2.1. TRANSCONDUCTANCE MULTIPLIER SYNAPSES
The equation of interest is that for the drain-source current, IDs, for a
MOSFET in the linear or triode region:IDS =
j.l.C ox W [
-1:--
(VGS - V T
)
VDS -
VDs2]
2--
(1)
Here, Cox is the oxide capacitance/area, j.l. the carrier mobility, W the
transistor gate width, L the transistor gate length, and VGS, VT, VDS the
transistor gate-source, threshold and drain-source voltages respectively.
.
. f
.
ThIs
expressIon
or IDS contams
a use f ul prod uct term:- j.l.CLox W x VGS X V .
DS
However, it also contains two other terms in V DS x VT and VDs2.
One approach might be to ignore this imperfection in the multiplication, in
the hope that the neural parallelism renders it irrelevant. We have chosen,
rather, to remove the unwanted terms via a second M OSFET, as shown in
Fig. 2.
787
788
Brownlow, Tarassenko, Murray, Hamilton, Han and Reekie
13 = 11-12
Figure 2 : Use of a second MOSFET to remove nonlinearities
(a transconductance multiplier).
The output current 13 is now given by:W1
13 = JJ.Cox [ L1 (VGSl - VT
W2
L 2 (VGS2 - V T
)
VDS1 -
)
VDS2
+
W 1 vDsl
L1
2
(2)
W 2 VDsL]
L2
2
The secret now is to select W 1, L 1, W 2, L 2, VGSb VGS2 , VDS1 and VDS2 to
cancel all terms except
W1
JJ.Cox L1 VGSl X VDS1
(3)
This is a fairly well-known circuit, and constitutes a Transconductance
Multiplier. It was reported initially for use in signal processing chips such
as filters [Denyer,1981 , Han,1984]. It would be feasible to use it directly in
a continuous time network, with analog voltages representing the {Sj}. We
choose to use it within a pulse-stream environment, to minimise the
uncertainty in determining the operating regime, and terminal voltages, of
the MOSFETs, as described above.
Fig. 3 shows two related pulse stream synapse based on this technique. The
presynaptic neural state Sj is represented by a stream of 0-5V digital,
asynchronous voltage pulses Vj ? These are used to switch a current sink and
source in and out of the synapse, either pouring current to a fixed voltage
node (excitation of the postsynaptic neuron), or removing it (inhibition).
The magnitude and direction of the resultant current pulses are determined
by the synapse weight, currently stored as a dynamic, analog voltage Tij.
Pulse-Firing Neural Chips for Hundreds of Neurons
(a)
(b)
State V J
Reference 1
r1
Reference V r
:r:
Tij
Vfixed
Vfixed
Referen~
Reference 3
I
Figure 3 : Use of a transconductance multiplier to
form fully programmable pulse-stream synapses.
The fixed voltage VJixed and the summation of the current pulses to give an
activity Xj = 'LTjjSj are both provided by an Operational Amplifier
integrator circuit, whose saturation characteristics incidentally apply a
sigmoid nonlinearity. The transistors Tl and T4 act as power supply
"on/off" switches in Fig. 3a, and in Fig 3b are replaced by a single
transistor, in the output "leg" of the synapse, Transistors T2 and T3 form the
transconductance multiplier. One of the transistors has the synapse voltage
Tij on its gate, the other a reference voltage, whose value determines the
crossover point between excitation and inhibition. The gate-source voltages
on T2 and T3 need to be substantially greater than the drain-source
voltages, to maintain linear operation. This is not a difficult constraint to
satisfy.
The attractions of these cells are that all the transistors are n-type, removing
the need for area-hungry isolation well structures, and In Fig. 3a, the
vertical line of drain-source connections is topologically attractive,
producing very compact layout, while Fig. 3b has fewer devices. It is not
yet clear which will prove optimal.
2.2. ASYNCHRONOUS "SWITCHED CAPACITOR" SYNAPSE
Fig. 4 shows a further variant, in the form of a "switched capacitor" pulse
stream synapse. Here the synapse voltage Tij is electrically buffered to
switched capacitor structure, clocked by the presynaptic neural pulse
waveforms. Packets of charge are therefore "metered out" to the current
integrator whose magnitude is controlled by Tij (positive or negative), and
789
790
Brownlow, Tarassenko, 1\1urray, Hamilton, Han and I{eekie
whose frequency by the presynaptic pulse rate. The overall principle is
therefore the same as that described for the transconductance multiplier
synapses, although the circuit level details are different.
-
Vj
Buffer
Vj
Integrator
Tij
~
T
I
X
1
/
-
I
Figure 4 : Asynchronous, "switched capacitor" pulse stream synapse.
Conventional synchronous switched capacitor techniques have been used in
neural integration [Tsividis,1987], but nowhere as directly as in this
example.
2.3. CHIP DETAILS AND RESULTS
Both the time-modulation and switched capacitor synapses have been tested
fully in silicon, and Fig. 5 shows a section of the time-modulation test chip.
This synapse currently occupies 174x73jl.m.
Figure 5 : Section, and single synapse, from time-modulation chip.
Pulse-Firing Neural Chips for Hundreds of Neurons
Three distinct pulse-stream synapse types have been presented, with
different operating schemes and characteristics. None has yet been used to
configure a large network, but this is now being done. Current estimates
for the number of synapses implementable using the two techniques
described above are as shown in Table 1, using an 8mmx8mm die as an
example.
The lack of direct scaling between transistor count and synapse count (e.g.
why does the factor 4111 not manifest itself as a much larger increase in
synapse count) can be explained. The raw number of transistors is not the
only factor in determining circuit area. Routing of power supplies, synapse
weight address lines, as well as storage capacitor size all take their toll, and
are common to both of the above synapse circuits. Furthermore, in analog
circuitry, transistors are almost certainly larger than minimum geometry,
and generally significantly larger, to minimise noise problems. This all gives
rise to a larger area than might be expected from simple arguments.
Clearly, however, we are in position to implement serious sized networks,
firstly with the time-modulation synapse, which is fully tested in silicon, and
later with the transconductance type, which is still under detailed design and
layout.
Table 1 : Estimated synapse count on 8mm die
SYNAPSE
Time modulation
Transconductance
Switched Capacitor
NO. OF
TRANSISTORS
ESTIMATED
NETWORK SIZE
11
4
4
= 6400 synapses
=15000 synapses
= 14000 synapses
In addition, we are developing new oscillator forms, techniques to
counteract leakage from dynamic nodes, novel inter-chip signalling strategies
specifically for pulse-stream systems, and non-volatile (a-Si) pulse stream
synapses. These are to be used for applications in text-speech synthesis,
pattern analysis and robotics. Details will be published as the work
progresses.
Acknowledgements
The authors are grateful to the UK Science and Engineering Research
Council, and the European Community (ESPRIT BRA) for its support of
this work. Dr. Han is grateful to the Korean Telecommunications
Authority, from whence he is on secondment in Edinburgh, and
KOSEF(Korea) for partial financial support.
791
792
Brownlow, Tarassenko, Murray, Hamilton, Han and Reekie
References
Daniell, 1989.
P. M. Daniell, W. A. J. Waller, and D. A. Bisset, "An
Implementation of Fully Analogue Sum-of-Product Neural Models,"
Proc. lEE Conf. on Artificial Neural Networks, pp. 52-56, ,1989.
Denyer ,1981.
P. B. Denyer and J. Mavor, "MOST Transconductance Multipliers for
Array Applications," lEE Proc. Pt. 1, vol. 128, no. 3, pp. 81-86, June
,1981.
EI-Leithy,1988.
N. EI-Leithy, M. Zaghloul, and R. W. Newcomb, "Implementation of
Pulse-Coded Neural Networks," Proc. 27th Conj. on Decision and
Control, pp. 334-336, ,1988.
Han,1984.
n S. Han and Song B. Park, "Voltage-Controlled Linear Resistors by
MaS Transistors and their Application to Active RC Filter MaS
Integration," Proc. IEEE, pp. 1655-1657, Nov., ,1984.
Murray,1987.
A. F. Murray and A. V. W. Smith, "Asynchronous Arithmetic for
VLSI Neural Systems," Electronics Letters, vol. 23, no. 12, pp. 642-3,
June, ,1987.
Murray,1988a.
A. F. Murray and A. V. W. Smith, "Asynchronous VLSI Neural
Networks using Pulse Stream Arithmetic," IEEE Journal of Solid-State
Circuits and Systems, vol. 23, no. 3, pp. 688-697, June, ,1988.
Murray,1988b.
A. F. Murray, L. Tarassenko, and A. Hamilton, "Programmable
Analogue Pulse-Firing Neural Networks," Neural Information
Processing Systems Conference, pp. 671-677, Morgan Kaufmann,
,1988.
Murray,1989a.
A. F. Murray, "Pulse Arithmetic in VLSI Neural Networks," IEEE
MICRO, vol. 9, no. 6, pp. 64-74, ,1989.
Murray,1989b.
A. F. Murray, A. Hamilton, H. M. Reekie, and L. Tarassenko,
"Pulse - Stream Arithmetic in Programmable Neural Networks," Int.
Symposium on Circuits and Systems, Portland, Oregon, pp. 1210-1212,
IEEE, ,1989.
Tsividis,1987.
Y. P. Tsividis and D. Anastassiou, "Switched - Capacitor Neural
Networks," Electronics Letters, vol. 23, no. 18, pp. 958 - 959, August,
,1987.
| 215 |@word cox:3 pulse:46 eng:3 solid:1 electronics:2 initial:1 contains:1 current:8 si:2 yet:2 remove:2 aside:1 fewer:1 device:2 unacceptably:1 signalling:1 accordingly:1 smith:2 authority:2 node:2 firstly:1 rc:1 direct:1 supply:2 symposium:1 prove:1 mayfield:1 inter:1 secret:1 expected:1 integrator:3 terminal:2 becomes:1 provided:1 circuit:12 substantially:1 developed:2 pseudo:2 act:1 charge:1 unwanted:1 esprit:1 uk:1 control:1 hamilton:7 producing:1 carrier:1 positive:1 engineering:1 waller:1 consequence:1 encoding:1 oxford:2 id:3 firing:8 path:1 modulation:6 might:2 ease:1 implement:2 area:5 crossover:1 significantly:1 projection:1 road:1 cannot:1 storage:3 context:1 equivalent:1 conventional:1 layout:2 attraction:1 array:1 financial:1 increment:1 pt:1 nowhere:1 tarassenko:7 electrical:3 region:1 predictable:1 environment:1 dynamic:4 grateful:2 upon:1 sink:1 easily:1 chip:13 emulate:1 represented:1 univ:2 mosfet:3 distinct:1 artificial:1 whose:4 larger:4 regeneration:1 itself:1 advantage:4 toll:1 transistor:15 product:1 lionel:1 r1:1 incidentally:1 cmos:2 exemplar:1 ij:1 progress:1 aesthetically:1 implies:1 direction:1 waveform:1 newcomb:1 functionality:1 filter:2 packet:1 occupies:1 routing:1 implementing:2 behaviour:1 niv:1 biological:2 summation:1 mm:1 great:1 circuitry:1 arrange:1 early:1 mosfets:1 proc:4 currently:2 council:1 hope:1 clearly:3 imperfection:1 rather:2 occupied:1 voltage:18 june:3 unsatisfactory:1 portland:1 whence:1 initially:2 vlsi:5 overall:1 integration:2 fairly:1 park:1 cancel:1 constitutes:1 t2:2 serious:1 micro:1 individual:2 replaced:1 geometry:1 fire:1 maintain:1 amplifier:1 interest:1 certainly:2 configure:1 partial:1 conj:1 korea:1 mobility:1 instance:1 hundred:6 reported:2 stored:1 lee:2 off:1 receiving:2 michael:1 vgs:5 synthesis:1 w1:2 choose:1 dr:1 conf:1 oxide:1 inefficient:1 nonlinearities:1 int:1 oregon:1 satisfy:1 stream:23 performed:1 later:3 doing:1 il:1 kaufmann:1 characteristic:3 t3:2 raw:1 none:1 brownlow:5 published:1 synapsis:14 frequency:3 pp:10 resultant:1 associated:1 proved:1 manifest:1 synapse:25 done:1 ox:1 furthermore:4 uct:1 clock:1 d:2 ei:3 ox1:1 lack:1 indicated:1 effect:1 multiplier:7 inspiration:1 chemical:1 volt:1 attractive:1 anastassiou:1 width:1 excitation:3 die:2 clocked:1 l1:3 hungry:1 novel:1 sigmoid:1 common:1 volatile:1 overview:1 leithy:3 pouring:1 insensitive:1 jl:1 discussed:1 analog:10 he:1 silicon:4 significant:2 buffered:1 nonlinearity:1 han:9 operating:4 inhibition:3 irrelevant:1 buffer:1 vt:3 seen:1 minimum:1 greater:1 morgan:1 bra:1 signal:3 ii:1 arithmetic:8 full:1 alan:1 smooth:1 coded:1 controlled:3 variant:1 essentially:1 vision:1 veo:1 robotics:1 cell:1 addition:1 source:7 w2:1 capacitor:9 viability:1 switch:2 xj:1 isolation:1 zaghloul:1 minimise:2 vgs2:2 whether:1 motivated:1 expression:1 synchronous:1 ul:1 song:2 render:1 speech:1 jj:2 programmable:5 ignored:1 tij:8 generally:2 clear:1 detailed:1 amount:1 estimated:2 vol:5 group:1 four:1 threshold:1 pj:1 sum:1 counteract:1 letter:2 uncertainty:1 telecommunication:2 topologically:1 family:1 announce:1 almost:1 architectural:2 decision:2 bisset:1 scaling:1 eh9:1 fl:2 expediency:1 alister:1 activity:4 constraint:1 argument:1 extremely:1 transconductance:9 martin:1 developing:1 electrically:1 postsynaptic:1 leg:1 explained:1 daniell:3 equation:3 remains:1 count:4 mechanism:1 sending:1 operation:1 apply:1 away:1 robustness:1 gate:5 triode:1 giving:1 murray:19 appreciate:1 leakage:1 move:1 capacitance:1 already:1 spike:1 strategy:1 detrimental:1 separate:1 vd:4 presynaptic:3 code:2 length:1 relationship:1 korean:2 difficult:2 negative:1 rise:2 design:1 implementation:2 unknown:1 perform:2 vertical:1 neuron:12 regenerate:1 implementable:1 august:1 community:1 connection:1 address:1 usually:1 dynamical:1 parallelism:1 pattern:1 regime:2 saturation:1 memory:1 analogue:2 power:2 hybrid:1 predicting:1 representing:1 scheme:1 altered:1 lij:4 text:1 l2:1 acknowledgement:1 drain:4 multiplication:1 determining:2 fully:8 interesting:1 digital:7 switched:8 reekie:5 principle:1 asynchronous:9 edinburgh:5 benefit:1 author:1 made:1 simplified:2 far:1 sj:8 nov:1 compact:2 ignore:1 active:1 xi:1 continuous:2 prod:1 why:2 table:2 operational:1 tsividis:3 investigated:1 european:1 vj:3 did:1 decrement:1 noise:3 body:1 fig:9 tl:1 clumsy:1 position:1 resistor:1 removing:2 denyer:3 magnitude:3 t4:1 easier:2 admitted:1 desire:1 determines:1 ma:2 sized:1 towards:1 oscillator:1 feasible:1 determined:3 except:1 specifically:1 pas:1 select:1 support:2 dept:3 tested:2 |
1,263 | 2,150 | Bayesian Monte Carlo
Carl Edward Rasmussen and Zoubin Ghahramani
Gatsby Computational Neuroscience Unit
University College London
17 Queen Square, London WC1N 3AR, England
edward,[email protected]
http://www.gatsby.ucl.ac.uk
Abstract
We investigate Bayesian alternatives to classical Monte Carlo methods
for evaluating integrals. Bayesian Monte Carlo (BMC) allows the incorporation of prior knowledge, such as smoothness of the integrand,
into the estimation. In a simple problem we show that this outperforms
any classical importance sampling method. We also attempt more challenging multidimensional integrals involved in computing marginal likelihoods of statistical models (a.k.a. partition functions and model evidences). We find that Bayesian Monte Carlo outperformed Annealed
Importance Sampling, although for very high dimensional problems or
problems with massive multimodality BMC may be less adequate. One
advantage of the Bayesian approach to Monte Carlo is that samples can
be drawn from any distribution. This allows for the possibility of active
design of sample points so as to maximise information gain.
1 Introduction
Inference in most interesting machine learning algorithms is not computationally tractable,
and is solved using approximations. This is particularly true for Bayesian models which
require evaluation of complex multidimensional integrals. Both analytical approximations,
such as the Laplace approximation and variational methods, and Monte Carlo methods
have recently been used widely for Bayesian machine learning problems. It is interesting
to note that Monte Carlo itself is a purely frequentist procedure [O?Hagan, 1987; MacKay,
1999]. This leads to several inconsistencies which we review below, outlined in a paper
by O?Hagan [1987] with the title ?Monte Carlo is Fundamentally Unsound?. We then
investigate Bayesian counterparts to the classical Monte Carlo.
Consider the evaluation of the integral:
(1)
where is
a probability (density), and
is the
we wish to integrate. For
function
example,
could
be
the
posterior
distribution
and
the
predictions
by a model
made
with parameters , or
could be the parameter prior and
the likelihood
so that equation (1) evaluates the marginal likelihood (evidence) for a model. Classical
Monte Carlo makes the approximation:
(2)
where
are random (not necessarily independent) draws
, which converges
to
. from
the right answer in the limit of large numbers
of
samples,
If
sampling
directly
from
is hard, or if high density regions in
do not match up with areas where
has large
magnitude,
it is also possible to draw samples from some importance
sampling distribution
to obtain the estimate:
(3)
As O?Hagan [1987] points out, there are two important
objections to these procedures.
not only depends on the values of
but also on the enFirst, the estimator
tirely
same set of samples
choice of the sampling distribution . Thus, if the
arbitrary
,
were obtained from
, conveying exactly the same information about
two different sampling distributions, two different estimates of would be obtained. This
dependence on irrelevant (ancillary)
information is unreasonable and violates the Likelihood Principle. The second
objection
is that classical Monte Carlo procedures entirely
ignore the values
of the
when forming the estimate. Consider the simple example of
three points
that are sampled from and the third happens to fall on the same point as the
second,
, conveying no extra information about the integrand. Simply aver-
aging the integrand at these three points, which is the classical Monte Carlo estimate, is
clearly inappropriate; it would make much more sense to average the first two (or the first
and third). In practice points are unlikely to fall on top of each other in continuous spaces,
however, a procedure that weights points equally regardless of their spatial distribution is
ignoring relevant information. To summarize the objections, classical Monte Carlo bases
its estimate on irrelevant information and throws away relevant information.
We seek to turn the problem of evaluating the integral (1) into a Bayesian inference problem
which, as we will see, avoids the inconsistencies of classical Monte Carlo and
can result
in better estimates. To do this, we think of the unknown desired quantity
as being
random. Although this interpretation is not the most usual one, it is entirely consistent with
the Bayesian view that all forms of uncertainty are represented
using
probabilities: in this
case uncertainty
arises because
we
afford to compute
at every location. Since
cannot
the desired
is a function
of
(which is unknown until we evaluate it) we proceed
to obtain the posterior over
by putting a prior on , combining it with the observations
which in turn implies a distribution over the desired .
A very convenient way of putting priors over functions is through Gaussian Processes (GP).
Under a GP prior
the joint distribution
of any (finite) number of function values (indexed
by the inputs, ) is Gaussian:
! # "
(4)
where here we take the mean to be zero. The covariance matrix is given by the covariance
function, a convenient choice being:1
"$ &
%('*)
+
$
:
,+.-/1032546 79
8
;
>: =
:
:
6
$
* < +
(5)
where the parameters are hyperparameters. Gaussian processes, including optimization
of hyperparameters, are discussed in detail in [Williams and Rasmussen, 1996].
1
Although the function values obtained are assumed to be noise-free, we added a tiny constant to
the diagonal of the covariance matrix to improve numerical conditioning.
2 The Bayesian Monte Carlo Method
The Bayesian Monte Carlo method starts with a prior
over the function,
and makes
inferences about from
giving the
a set
of samples
posterior distribution
. Under a GP prior the posterior is (an infinite dimensional
joint)
since the
Gaussian;
integral
eq. (1) is just a linear projection (on the direction defined
by
), the posterior
is also Gaussian, and fully characterized by its mean and
variance. The average over functions of eq. (1) is the expectation of the average function:
where
(6)
is the posterior mean function. Similarly, for the variance:
%('*)
6
6
6
%('*) 4
=
(7)
where
is the posterior covariance. The standard results for the GP model for the
=
posterior mean and covariance
are:
"! $#
#
"
&%
and
%('*) 4
6
'!
"
$#
(!
)%
!
*#
(8)
where and are the observed inputs and function values respectively. In general combining eq. (8) with eq. (6-7) may lead to expressions which are difficult to evaluate, but there
are several interesting special cases.
If the density
and thecovariance
eq. (5) are both Gaussian, we obtain ana
,+ .function
-
and the Gaussian kernels on the data points are
lytical results. In detail, if
*/
10 diag
+ +
then the expectation evaluates to:
'2
"
%
,+ - 8
32
0
%
%
-5476
/1032 6
.8
:9
*/
6
;+
*0<4<-
%
6
*/
;+
(9)
a result which has previously been derived under the name of Bayes-Hermite Quadrature
[O?Hagan, 1991]. For the variance, we get:
+.- 7
7=
=
0
%
->4&6?= %
=
6 "
$8
32
%
2
(10)
2
lead to analytical results include polynomial
with as defined in eq. (9). Other choices
that
kernels and mixtures of Gaussians for
.
2.1 A Simple Example
To illustrate the method we evaluated the integral of a one-dimensional function
under a
Gaussian
density
(figure
1,
left).
We
generated
samples
independently
from
, evalu
ated
at those points, and optimised the hyperparameters of our Gaussian process fit
to the function. Figure 1 (middle) compares the error in the Bayesian Monte Carlo (BMC)
estimate of the integral (1) to the Simple Monte Carlo (SMC) estimate using the same samples.
we would
expect the squared error in the Simple Monte Carlo estimate decreases
is the
< Aswhere
sample size. In contrast, for more than about 10 samples, the
as
BMC
estimate improves at a much higher rate. This is achieved because the prior on allows
the method to interpolate between sample points.
Moreover, whereas the SMC estimate is
invariant to permutations of the values on the axis, BMC makes use of the smoothness of
the function. Therefore, a point in a sparse region is far more informative about the shape
of the function for BMC than points in already densely sampled areas. In SMC if two samples happen to fall close to each other the function value there will be counted with double
weight.
This effect means that large numbers of samples are needed to adequately represent
.
BMC
circumvents this problem by analytically integrating its mean function w.r.t.
.
In figure 1 left, the negative log density of the true value of the integral under the predictive distributions are compared for BMC and SMC. For not too small sample sizes, BMC
outperforms SMC. Notice however, that for very small sample sizes BMC occasionally
has
very bad performance.
This
is
due
to
examples
where
the
random
draws
of
lead
to
func
tion values
that are consistent with much longer length scale than the true function;
the mean prediction becomes somewhat inaccurate, but worse still, the inferred variance
becomes very small (because a very slowly varying function is inferred), leading to very
poor performance compared to SMC. This problem is to a large extent caused by the optimization of the length scale hyperparameters of the covariance function; we ought instead
to have integrated over all possible length scales. This integration would effectively ?blend
in? distributions with much larger variance (since the data is also consistent with a shorter
length scale), thus alleviating the problem, but unfortunately this is not possible in closed
form. The problem disappears for sample sizes of around 16 or greater.
In the previous example, we chose
to be Gaussian. If you wish to use BMC to integrate
w.r.t. non-Gaussian densities then an importance re-weighting trick becomes necessary:
<
(11)
and is a Gaussian and
where the Gaussian process models
is
an arbitrary density
which
can
be
evaluated.
See
Kennedy [1998] for extension to non
Gaussian
.
2.2 Optimal Importance Sampler
For the simple example discussed above, it is also interesting to ask whether the efficiency
of SMC could be improved by generating
independent samples from more-cleverly designed distributions.
As
we
have
seen
in
equation
(3), importance sampling gives an unbi
ased estimate of by sampling
from
and computing:
where
wherever
(12)
. The variance of this estimator is given by:
56
(13)
Using calculus of variations it is simple to show that the optimal (minimum variance) importance sampling distribution is:
(14)
which we can substitute into equation (13) to get
the minimum variance, . If
is
always non-negative
or
non-positive
then
,
which
is
unsurprising
given
that
we
needed to know in advance to normalise . For functions that take on both positive and
?2
10
Bayesian inference
Simple Monte Carlo
Optimal importance
0.4
?3
10
average squared error
0.3
0.2
0.1
0
?0.1
?0.2
?4
10
?5
10
?6
10
?0.3
?0.4
?0.5
?4
Bayesian inference
Simple Monte Carlo
minus log density of correct value
function f(x)
measure p(x)
0.5
20
15
10
5
0
?5
?7
?2
0
2
10
4
1
10
2
10
sample size
1
10
2
10
sample size
Figure 1: Left: a simple one-dimensional function
(full) and Gaussian density (dashed)
with respect to which we wish to integrate . Middle: average squared error for simple Monte Carlo sampling from (dashed), the optimal achievable bound for importance
sampling (dot-dashed), and the Bayesian Monte Carlo estimates. The values plotted are
averages over up to 2048 repetitions. Right: Minus the log of the Gaussian predictive density with mean eq. (6) and variance eq. (7), evaluated at the true value of the integral (found
by numerical integration), ?x?. Similarly for the Simple Monte Carlo procedure, where the
mean and variance of the predictive distribution are computed from the samples, ?o?.
<
6
negative values
which is a constant times the variance of
a Bernoulli random variable (sign
). The lower bound from this optimal importance
sampler as a function of number of samples is shown in figure 1, middle. As we can
see, Bayesian Monte Carlo improves on the optimal importance sampler considerably. We
stress that the optimal importance sampler is not practically achievable since it requires
knowledge of the quantity we are trying to estimate.
3 Computing Marginal Likelihoods
We now consider the problem of estimating the marginal likelihood of a statistical model.
This problem is notoriously difficult and very important, since it allows for comparison of
different models. In the physics literature it is known as free-energy estimation. Here we
compare the Bayesian Monte Carlo method to two other techniques: Simple Monte Carlo
sampling (SMC) and Annealed Importance Sampling (AIS).
Simple Monte Carlo, sampling from the prior, is generally considered inadequate for this
problem, because the likelihood is typically sharply peaked and samples from the prior are
unlikely to fall in these confined areas, leading to huge variance in the estimates (although
they are unbiased). A family of promising ?thermodynamic integration? techniques for
computing marginal likelihoods are discussed under the name of Bridge and Path sampling
in [Gelman and Meng, 1998] and Annealed Importance Sampling (AIS) in [Neal, 2001].
The central idea is to divide one difficult integral into a series of easier ones, parameterised
by (inverse) temperature, . In detail:
-
-
%
where
and
is the inverse temperature of the annealing schedule and
where
To compute
each fraction we sample from equilibrium from the distribution
%
and compute importance weights:
,!
%
(15)
!"
!
%
%
;
%
%
%
.
(16)
In practice can be set to 1, to allow very slow reduction in temperature. Each of the
intermediate ratios are much easier to compute than the original ratio, since the likelihood
function to the power of a small number is much better behaved that the likelihood itself.
Often elaborate non-linear cooling schedules are used, but for simplicity we will just take
a linear schedule for the inverse temperature. The samples at each temperature are drawn
using a single Metropolis proposal, where the proposal width is chosen to get a fairly high
fraction of acceptances.
The model in question for which we attempt to compute the marginal likelihood was itself a Gaussian process regression fit to the an artificial dataset suggested
+ - by [Friedman,
1988].2 We had 9 length scale hyperparameters, a signal variance ( ) and an explicit
noise variance parameter. Thus the marginal likelihood is an integral
a 7 dimensional
over
priors.
hyperparameter space. The log of the hyperparameters are given
Figure 2 shows a comparison of the three methods. Perhaps surprisingly, AIS and SMC are
seen to be very comparable, which can be due to several reasons: 1) whereas the SMC samples are drawn independently, the AIS samples have considerable auto-correlation because
of the Metropolis generation mechanism, which hampers performance for low sample sizes,
2) the annealing schedule was not optimized nor the proposal width adjusted with temperature, which might possibly have sped up convergence. Further, the difference between
AIS and SMC would be more dramatic in higher dimensions and for more highly peaked
likelihood functions (i.e. more data).
The Bayesian Monte Carlo method was run on the same samples as were generate
by
the
AIS procedure. Note that BMC can use samples from any distribution, as long as
can
be evaluated. Another obvious choice for generating samples for BMC would be to use
an MCMC method to draw samples from the posterior. Because BMC needs to model the
integrand using a GP, we need to limit the number of samples
computation (for fitting
. Thussince
hyperparameters
and
computing
the
?s)
scales
as
for
sample
greater than
7 we limit the number of samples to 7 , chosen equally spaced fromsize
the AIS Markov
chain. Despite this thinning of the samples we see a generally superior performance of
BMC, especially for smaller sample sizes. In fact, BMC seems to perform equally well for
almost any of the investigated sample sizes. Even for this fairly large number of samples,
the generation of points from the AIS still dominates compute time.
4 Discussion
An important aspect which we have not explored in this paper is the idea that the GP model
used to fit the integrand gives errorbars (uncertainties) on the integrand. These error bars
2
was 100 samples generated from the 5-dimensional function "!$#&%('(''(%)!+*&,./&021)35The
4 "67! data
# !98&,;:=< 0 "!
>@? 0 ' AB, 8 : /&0 !9C:=A! * :=D , where D is zero mean unit variance Gaussian
noise and the inputs are sampled independently from a uniform [0, 1] distribution.
Log Marginal Likelihood
?45
?50
?55
?60
True
SMC
AIS
BMC
?65
?70
3
10
4
10
Number of Samples
5
10
Figure 2: Estimates of the marginal likelihood for different sample sizes using Simple
Monte Carlo sampling (SMC; circles, dotted line), Annealed Importance Sampling (AIS;
, dashed line), and Bayesian Monte Carlo (BMC;
triangles, solid line). The true value
sample
(solid straight line) is estimated 6
from
a
single
long run of AIS. For comparison,
(which is an upper bound on the true value).
the maximum log likelihood is
could be used to conduct an experimental design,
i.e. active learning. A simple approach
would
and
be to evaluate the function at points where the GP has large uncertainty
is not too small:
the
expected contribution to the uncertainty in the estimate of the
. For a fixed Gaussian Process covariance function these design
integral scales as
points can often be pre-computed, see e.g. [Minka, 2000]. However, as we are adapting the
covariance function depending on the observed function values, active learning would have
to be an integral part of the procedure. Classical Monte Carlo approaches cannot make use
of active learning since the samples need to be drawn from a given distribution.
When using BMC to compute marginal likelihoods, the Gaussian covariance function used
here (equation 5) is not ideally suited to modeling the likelihood. Firstly, likelihoods are
non-negative whereas the prior is not restricted in the values the function can take. Secondly, the likelihood tends to have some regions of high magnitude and variability and
other regions which are low and flat; this is not well-modelled by a stationary covariance
function. In practice this misfit between the GP prior and the function modelled has even
occasionally led to negative values for the estimate of the marginal likelihood! There could
be several approaches to improving the appropriateness of the prior. An importance distribution such as one computed from a Laplace approximation or a mixture of Gaussians
can be used to dampen the variability in the integrand [Kennedy, 1998]. The GP could be
used to model the log of the likelihood [Rasmussen, 2002]; however this makes integration
more difficult.
The BMC method outlined in this paper can be extended in several ways. Although the
choice of Gaussian process priors is computationally convenient in certain circumstances,
in general other function approximation priors can be used to model the integrand. For
discrete (or mixed) variables the GP model could still be used with appropriate choice of
covariance function. However, the resulting sum (analogous to equation 1) may be difficult
to evaluate. For discrete , GPs are not directly applicable.
Although BMC has proven successful on the problems presented here, there are several
limitations to the approach. High dimensional integrands can prove difficult to model. In
such cases a large number of samples may be required to obtain good estimates of the
function. Inference using a Gaussian Process prior is at present limited computationally
to a few thousand samples. Further, models such as neural networks and mixture models
exhibit an exponentially large number of symmetrical modes in the posterior. Again modelling this with a GPprior
would typically be difficult. Finally, the BMC method requires
that the distribution
can be evaluated. This contrasts with classical MC
many
where
methods only require that samples can be drawn from some distribution
, for which
the normalising constant is not necessarily known (such as in equation 16). Unfortunately,
this limitation makes it difficult, for example, to design a Bayesian analogue to Annealed
Importance Sampling.
We believe that the problem of computing an integral using a limited number of function
evaluations should be treated as an inference problem and that all prior knowledge about
the function being integrated should be incorporated into the inference. Despite the limitations outlined above, Bayesian Monte Carlo makes it possible to do this inference and
can achieve performance equivalent to state-of-the-art classical methods despite using a
fraction of sample evaluations, even sometimes exceeding the theoretically optimal performance of some classical methods.
Acknowledgments
We would like to thank Radford Neal for inspiring discussions.
References
Friedman, J. (1988). Multivariate Adaptive Regression Splines. Technical Report No. 102, November 1988, Laboratory for Computational Statistics, Department of Statistics, Stanford University.
Kennedy, M. (1998). Bayesian quadrature with non-normal approximating functions, Statistics and
Computing, 8, pp. 365?375.
MacKay, D. J. C. (1999). Introduction to Monte Carlo methods. In Learning in Graphical Models,
M. I. Jordan (ed), MIT Press, 1999.
Gelman, A. and Meng, X.-L. (1998) Simulating normalizing constants: From importance sampling
to bridge sampling to path sampling, Statistical Science, vol. 13, pp. 163?185.
Minka, T. P. (2000) Deriving quadrature rules from Gaussian processes, Technical Report, Statistics
Department, Carnegie Mellon University.
Neal, R. M. (2001). Annealed Importance Sampling, Statistics and Computing, 11, pp. 125?139.
O?Hagan, A. (1987). Monte Carlo is fundamentally unsound, The Statistician, 36, pp. 247-249.
O?Hagan, A. (1991). Bayes-Hermite Quadrature, Journal of Statistical Planning and Inference, 29,
pp. 245?260.
O?Hagan, A. (1992). Some Bayesian Numerical Analysis. Bayesian Statistics 4 (J. M. Bernardo,
J. O. Berger, A. P. Dawid and A. F. M. Smith, eds), Oxford University Press, pp. 345?365 (with
discussion).
C. E. Rasmussen (2003). Gaussian Processes to Speed up Hybrid Monte Carlo for Expensive
Bayesian Integrals, Bayesian Statistics 7 (J. M. Bernardo, M. J. Bayarri, J. O. Berger, A. P. Dawid,
D. Heckerman, A. F. M. Smith and M. West, eds), Oxford University Press.
Williams, C. K. I. and C. E. Rasmussen (1996). Gaussian Processes for Regression, in D. S. Touretzky, M. C. Mozer and M. E. Hasselmo (editors), NIPS 8, MIT Press.
| 2150 |@word middle:3 achievable:2 polynomial:1 seems:1 calculus:1 seek:1 covariance:12 dramatic:1 minus:2 solid:2 reduction:1 series:1 outperforms:2 numerical:3 happen:1 partition:1 informative:1 shape:1 designed:1 stationary:1 smith:2 normalising:1 location:1 firstly:1 hermite:2 prove:1 fitting:1 multimodality:1 theoretically:1 expected:1 nor:1 planning:1 inappropriate:1 becomes:3 estimating:1 moreover:1 ought:1 every:1 multidimensional:2 bernardo:2 exactly:1 uk:2 unit:2 positive:2 maximise:1 tends:1 limit:3 aging:1 despite:3 oxford:2 meng:2 optimised:1 path:2 might:1 chose:1 challenging:1 limited:2 smc:13 acknowledgment:1 practice:3 procedure:7 area:3 adapting:1 convenient:3 projection:1 pre:1 integrating:1 zoubin:2 get:3 cannot:2 close:1 gelman:2 www:1 equivalent:1 annealed:6 williams:2 regardless:1 independently:3 simplicity:1 estimator:2 rule:1 deriving:1 variation:1 laplace:2 analogous:1 massive:1 alleviating:1 carl:1 gps:1 trick:1 dawid:2 expensive:1 particularly:1 hagan:7 cooling:1 observed:2 solved:1 thousand:1 region:4 decrease:1 mozer:1 ideally:1 predictive:3 purely:1 efficiency:1 triangle:1 joint:2 represented:1 london:2 monte:36 artificial:1 widely:1 larger:1 stanford:1 statistic:7 think:1 gp:11 itself:3 advantage:1 analytical:2 ucl:2 relevant:2 combining:2 achieve:1 convergence:1 double:1 generating:2 converges:1 illustrate:1 depending:1 ac:2 eq:8 edward:2 throw:1 implies:1 appropriateness:1 direction:1 correct:1 ancillary:1 ana:1 violates:1 require:2 secondly:1 adjusted:1 extension:1 practically:1 around:1 considered:1 normal:1 equilibrium:1 estimation:2 outperformed:1 applicable:1 title:1 bridge:2 hasselmo:1 repetition:1 dampen:1 mit:2 clearly:1 gaussian:25 always:1 varying:1 derived:1 bernoulli:1 likelihood:22 modelling:1 contrast:2 sense:1 inference:10 inaccurate:1 unlikely:2 integrated:2 typically:2 spatial:1 special:1 mackay:2 integration:4 marginal:11 fairly:2 art:1 integrands:1 sampling:23 bmc:22 peaked:2 report:2 spline:1 fundamentally:2 few:1 unsound:2 densely:1 interpolate:1 hamper:1 statistician:1 attempt:2 friedman:2 ab:1 huge:1 acceptance:1 investigate:2 possibility:1 highly:1 evaluation:4 mixture:3 wc1n:1 chain:1 integral:16 necessary:1 shorter:1 indexed:1 conduct:1 divide:1 desired:3 re:1 plotted:1 circle:1 modeling:1 ar:1 queen:1 uniform:1 successful:1 inadequate:1 too:2 unsurprising:1 answer:1 considerably:1 density:10 physic:1 squared:3 central:1 again:1 slowly:1 possibly:1 worse:1 leading:2 caused:1 depends:1 ated:1 tion:1 view:1 closed:1 bayarri:1 start:1 bayes:2 contribution:1 square:1 variance:15 spaced:1 conveying:2 misfit:1 modelled:2 bayesian:27 mc:1 carlo:36 notoriously:1 kennedy:3 straight:1 touretzky:1 ed:3 evaluates:2 energy:1 pp:6 involved:1 minka:2 obvious:1 gain:1 sampled:3 dataset:1 ask:1 knowledge:3 improves:2 schedule:4 thinning:1 higher:2 improved:1 evaluated:5 just:2 parameterised:1 until:1 correlation:1 mode:1 perhaps:1 behaved:1 believe:1 name:2 effect:1 true:7 unbiased:1 counterpart:1 adequately:1 analytically:1 laboratory:1 neal:3 width:2 trying:1 stress:1 temperature:6 variational:1 recently:1 superior:1 sped:1 conditioning:1 exponentially:1 discussed:3 interpretation:1 mellon:1 ai:11 smoothness:2 outlined:3 similarly:2 had:1 dot:1 longer:1 base:1 posterior:10 multivariate:1 irrelevant:2 occasionally:2 certain:1 inconsistency:2 seen:2 minimum:2 greater:2 somewhat:1 dashed:4 signal:1 full:1 thermodynamic:1 technical:2 match:1 england:1 characterized:1 long:2 equally:3 prediction:2 regression:3 circumstance:1 expectation:2 kernel:2 represent:1 sometimes:1 achieved:1 confined:1 proposal:3 whereas:3 objection:3 annealing:2 extra:1 jordan:1 intermediate:1 fit:3 ased:1 idea:2 whether:1 expression:1 proceed:1 afford:1 adequate:1 generally:2 inspiring:1 http:1 generate:1 notice:1 dotted:1 sign:1 neuroscience:1 estimated:1 discrete:2 hyperparameter:1 carnegie:1 vol:1 putting:2 drawn:5 fraction:3 sum:1 run:2 inverse:3 uncertainty:5 you:1 family:1 almost:1 draw:4 circumvents:1 comparable:1 entirely:2 bound:3 incorporation:1 sharply:1 flat:1 integrand:8 aspect:1 speed:1 department:2 poor:1 cleverly:1 smaller:1 heckerman:1 metropolis:2 wherever:1 happens:1 invariant:1 restricted:1 computationally:3 equation:6 previously:1 turn:2 mechanism:1 needed:2 know:1 tractable:1 evalu:1 gaussians:2 unreasonable:1 away:1 appropriate:1 simulating:1 frequentist:1 alternative:1 substitute:1 original:1 top:1 include:1 graphical:1 giving:1 ghahramani:1 especially:1 approximating:1 classical:12 added:1 quantity:2 already:1 blend:1 question:1 dependence:1 usual:1 diagonal:1 exhibit:1 thank:1 normalise:1 extent:1 reason:1 length:5 berger:2 ratio:2 difficult:8 unfortunately:2 negative:5 design:4 unknown:2 perform:1 upper:1 observation:1 markov:1 finite:1 november:1 extended:1 variability:2 incorporated:1 arbitrary:2 inferred:2 required:1 optimized:1 errorbars:1 nip:1 suggested:1 bar:1 below:1 summarize:1 including:1 analogue:1 power:1 treated:1 hybrid:1 improve:1 disappears:1 axis:1 auto:1 func:1 prior:19 review:1 literature:1 fully:1 expect:1 permutation:1 mixed:1 interesting:4 generation:2 limitation:3 proven:1 integrate:3 consistent:3 principle:1 editor:1 tiny:1 surprisingly:1 rasmussen:5 free:2 allow:1 fall:4 sparse:1 dimension:1 evaluating:2 avoids:1 made:1 adaptive:1 counted:1 far:1 ignore:1 active:4 symmetrical:1 assumed:1 continuous:1 promising:1 ignoring:1 improving:1 investigated:1 complex:1 necessarily:2 diag:1 noise:3 hyperparameters:7 quadrature:4 west:1 elaborate:1 gatsby:3 slow:1 wish:3 explicit:1 exceeding:1 third:2 weighting:1 aver:1 bad:1 explored:1 evidence:2 dominates:1 normalizing:1 effectively:1 importance:20 magnitude:2 easier:2 suited:1 led:1 simply:1 forming:1 radford:1 considerable:1 hard:1 infinite:1 sampler:4 experimental:1 college:1 arises:1 evaluate:4 mcmc:1 |
1,264 | 2,151 | Discriminative Binaural Sound Localization
Ehud Ben-Reuven and Yoram Singer
School of Computer Science & Engineering
The Hebrew University, Jerusalem 91904, Israel
[email protected], [email protected]
Abstract
Time difference of arrival (TDOA) is commonly used to estimate the azimuth of a source in a microphone array. The most common methods
to estimate TDOA are based on finding extrema in generalized crosscorrelation waveforms. In this paper we apply microphone array techniques to a manikin head. By considering the entire cross-correlation
waveform we achieve azimuth prediction accuracy that exceeds extrema
locating methods. We do so by quantizing the azimuthal angle and
treating the prediction problem as a multiclass categorization task. We
demonstrate the merits of our approach by evaluating the various approaches on Sony?s AIBO robot.
1 Introduction
In this paper we describe and evaluate several algorithms to perform sound localization in
a commercial entertainment robot. The physical system being investigated is composed of
a manikin head equipped with a two microphones and placed on a manikin body. This type
of systems is commonly used to model sound localization in biological systems and the
algorithms used to analyze the signal are usually inspired from neurology. In the case of an
entertainment robot there is no need to be limited to a neurologically inspired model and
we will use combination of techniques that are commonly used in microphone arrays and
statistical learning. The focus of the work is the task of localizing an unknown stationary
source (compact in location and broad in spectrum). The goal is to find the azimuth angle
of the source relative to the head.
A common paradigm to approximately find the location of a sound source employs a microphone array and estimates time differences of arrival (TDOA) between microphones
in the array (see for instance [1]). In a dual-microphone array it is usually assumed that
the difference in the two channels is limited to a small time delay (or linear phase in frequency domain) and therefore the cross-correlation is peaked at the the time corresponding
to the delay. Thus, methods that search for extrema in cross-correlation waveforms are
commonly used [2]. The time delay approach is based on the assumption that the sound
waves propagate along a single path from the source to the microphone and that the microphone response of the two channels for the given source location is approximately the
same. In order for this to hold, the microphones should be identical, co-aligned, and, near
each other relative to the source. In addition there should not be any obstructions between
or near the microphones. The time delay assumption fails in the case of a manikin head: the
microphone are antipodal and in addition the manikin head and body affect the response
in a complex way. In our system the distance to the supporting floor was also significant.
Our approach for overcoming these difficulties is composed of two stages. First, we perform signal processing based on the generalized cross correlation transform called Phase
Transform (PHAT) also called Cross Power Spectrum Phase (CPSP). This signal processing removes to a large extent variations due the sound source. Then, rather than proceeding
with peak-finding we employ discriminative learning methods by casting the azimuth estimation as a multiclass prediction problem. The results achieved by combining the two
stages gave improved results in our experimental setup.
This paper is organized as follows. In Sec. 2 we describe how the signal received in the
two microphones was processed to generate accurate features. In Sec. 3 we outline the
supervised learning algorithm we used. We then discuss in Sec. 4 approaches to combined predictions from multiple segments. We describe experimental results in Sec. 5 and
conclude with a brief discussion in Sec. 6.
2 Signal Processing
Throughout the paper we denote signals in the time domain by lower case letters and in the
frequency domain by upper case letters. We denote the convolution operator between two
signals by and the correlation operator by . The unknown source signal is denoted by
and thus its spectrum is . The source signal passes through different physical setup and is
received at the right and left microphones. We denote the received signals by and . We
model the different physical media, the signal passes through, as two linear systems whose
frequency response is denoted by and . In addition the signals are contaminated with
noise that may account for non-linear effects such as room reverberations (see for instance
[3] for more detailed noise models). Thus, the received signals can be written in the time
and frequency domain as,
(1)
(2)
Since the source signal is typically non-stationary we break each training and test signal into segments and perform the processing described in the sequel based on short-time
Fourier transform. Let be the number of segments a signal is divided into and the
number of samples in a single segment. Each is multiplied by a Hanning window and
padded with zeros to smooth the end-of-segment effects and increase the resolution of the
short-time Fourier transform (see for instance [8]). Denote by and the left and right
signal-segments after the above processing. Based on the properties of the Fourier transform, the local cross-correlation between the two signals can be computed efficiently by
the inverse Fourier transform, denoted , of the product of the spectrum of and the
complex conjugate of the spectrum of ,
! #" %
$
&
(3)
Had the difference between the two signals been a mere time delay due to the different
location of the microphones, the cross correlation would have obtained its maximal value
at a point which corresponds to the time-lag between the received signals. However, since
the source signal passes through different physical media the short-time cross-correlation
does not necessarily obtain a large value at the time-lag index. It is therefore common (see
for instance [1]) to multiply the spectrum of the cross-correlation by a weighting function
in order to compensate for the differences in the frequency responses obtained at the two
microphones. Denoting the spectral shaping function for the ' th segment by ( , the generalization cross-correlation from Eq. (3) is, !
)*
+
" ( $ & . For
?plain? cross-correlation, ( ,.-0/ is equal to 1 at each (discrete) frequency - . In our tests
we found that a globally-equalized cross-correlation gives better results. The transform is
obtained by setting, ( ,.-0/ 1325476 where 486 is the average over all measurements and both
channels of 9 ,:-0/ 9 ; . Finally, for PHAT the weight for the spectral point - is,
1
( ,.-0/
:
,
<
/
9
$ ,:-</ 9
To further motivate and explain the PHAT weighting scheme, we build on the derivation
in [5] and expand the PHAT assuming that the noise is zero. In PHAT the spectral value at
frequency point - (prior to the inverse Fourier transform) is,
( ,.-0/ ,.-0/ $ ,:-</
., -0/ $ ., -0/
9 ., -0/ $ ., -0/ 9
(4)
Inserting Eq. (1) and Eq. (2) into Eq. (4) without noise we get,
(
,:-0/ ,.-0/ $ ,.-0/
:, -</ ,:-</ $ ., -0/ $ ., -0/
- / ., -0/ $ ., -0/ $ ., -0/ 9
9 ,.0
Therefore, assuming the noise is zero, PHAT eliminates the
contribution of the unknown source and the entire waveform of PHAT is only a function of the physical setup. If all
other physical parameters are constant, the PHAT waveform
(as well as its peak location) is a function of the azimuth angle of the sound source relative to the manikin head. This
is of course an approximation and the presence of noise and
changes in the environment result in a waveform that deviates from the closed-form given in Eq. (5). In Fig. 1 we show
the empirical average of the waveform for PHAT and for the
equalized cross-correlation, the vertical bars represent an error of 1 . In both cases, the location of the maximal correlation is clearly at as expected. Nonetheless, the high variance, especially in the case of the equalized cross-correlation
imply that classification of individual segments may often be
rather difficult.
$
9 9:9 9
(5)
0.6
0.5
0.4
0.3
0.2
0.1
0
?0.1
?5
?4
?3
?2
?1
0
1
2
3
4
5
0
1
2
3
4
5
0.05
0.04
0.03
0.02
0.01
0
?0.01
?0.02
?0.03
?5
?4
?3
?2
?1
Figure 1: Average wave-
form with standard deviIn practice, we found that it suffices to take only the ener
afation for
getic portion of the generalized cross-correlation waveforms
ter performing PHAT (top)
by considering only time lags of
through
samples. In
and the equalized crosswhat follows we will take this part to be the waveform. Forcorrelation (bottom).
mally, the feature vector of the ' th segment is defined as,
, ! ,
/
! ,
/ /
(6)
87
were
was set to be bigger than the maximal lag in samples between the two channels,
2 where is the head diameter and is speed of sound.
Summarising, the signal processing we perform is based on short time Fourier transform
of the signals received at the two microphones. From the two spectrums we then compute
the generalized cross-correlation using one of the three weighting schemes described above
and taking only
*1 samples of the resulting waveforms as the feature vectors. We now
move our focus to classification of a single segment.
3 Single Segment Classification
Traditional approaches to sound localization search for the the position of the extreme value
in the generalized cross-correlation waveform that were derived in Sec. 2. While being
intuitive, this approach is prone to noise. Peak location can be considered as a reduction
in dimensionality, from
1 to 1 , of the feature vectors , however we have shown
in Eq. 5 that the entire waveform of PHAT can be used as a feature vector to localise the
source. Indeed, in Sec. 5 we report experimental results which show that peak-finding is
significantly inferior to methods that we now describe, that uses the entire waveform. In all
techniques, peak-location and waveform, we used supervised learning to build a model of
the data using a training set and then used a test set to evaluate the learned model.
In a supervised learning setting, we have access to labelled examples and the goal is to
find a mapping from the instance domain (the peak-location or waveforms in our setting)
to a response variable (the azimuth angle). Since the angle is a continuous variable the
first approach that comes to mind is using a linear or non-linear regressors. However,
we found that regression algorithms such as Widrow-Hoff [10] yielded inferior results.
Instead of treating the learning problem as a regression problem, we quantized the angle and
converted the sound localization problem into a multiclass decision problem. Formally, we
where
into non-overlapping intervals
bisected the interval
78
/
,
/
<
2
<
2 and
1 72 1 . We now can transform the
1
! where
real-valued angle of the ' th segment, , into a discrete variable
78
%& . After this quantization, the training set is composed of instance-label
" iff $
#
pairs '
)(+ * and the first task is to find a classification rule from the peak-location or
, . We will first describe the method used for peak-location
waveforms space into 1
77
and then we will describe two discriminative methods to classify the waveform. The first
is based on a multiclass version of the Fisher linear discriminant [7] and is very simple to
implement. The second employs recent advances in statistical learning and can be used in
an online fashion. It thus can cope, to some extent with changes, in the environment such
as moving elements that change the reverberation properties of the physical media.
Peak location classification: Due to the relative low sampling frequency (-/. 10 " 21 )
spline interpolation was used to improve the peak location. In microphone arrays it is common to translate the peak-location to an estimate of the source azimuth using a geometric
formula. However, this was found to be inappropriate due to the internal reverberations
generated by the manikin head. We thus used the classification method describe in [4].
The peak location data was modelled using a separate histogram for each direction " . For
a given direction
%& , all the training measurements for which
%& are used to
!
, ! *1 / /
3 6
build a single histogram: , ! 9 " / 4
where ;< is 1 if
5 798:*
%
; is true and otherwise, is the size of the bin in the histogram, = >?; =
, and 9 9
is the number of bins. An estimate of the probability density function was taken to be
A
! B !
, ! *1 / / 9 " / 25
the normalized histogram step function: @ , 9 " / , C
%
" .
where
is the number of training measurements for which #
%
In order to classify new test data we simply compute the likelihood of the observed measurement under each distribution and choose the class attaining the maximal likelihood
(ML) score with respect to the distribution defined by the histogram,
A
D+EFHGID:J @
@
%
, 9" /
(7)
Multiclass Fisher discriminant: Generalising the Fisher discriminant for binary classification problems to multiclass settings, each class is modelled as a multivariate normal
distribution. To do so we divide the training set into subsets where the " th subset corresponds to measurements from azimuth in
%& . The density function of the " th class is
A
, 9" /
K
, :;
1
/ 9L
%
9+M
JNPO
1
, R
Q
%
/ S L %
,
Q
% U/ T
where S is the transpose of , '
1 is its dimensionality, Q denotes the mean of
%
the normal distribution, and L
the covariance matrix. Each mean and covariance matrix
%
are set to be the maximum likelihood estimates,
Q@ %
1
%W5 V 7
8* %
YX @L %
1
%
1 65 7
V
8* %
,
Q @ % /
S,
New test waveforms were then classified using the ML formula, Eq. 7.
Q@ %
/
The advantage of Fisher linear discriminant is that it is simple and easy to implement.
However, it degenerates if the training data is non-stationary, as often is the case in sound
localization problems due to effects such as moving objects. We therefore also designed,
implemented and tested a second discriminative methods based on the Perceptron.
Online Learning using Multiclass Perceptron with Kernels: Despite, or because of,
its age the Perceptron algorithm [9] is a simple and effective algorithm for classification.
We chose the Perceptron algorithm for its simplicity, adaptability, and ease in incorporating
Mercer kernels described below. The Perceptron algorithm is a conservative online algorithm: it receives an instance, outputs a prediction for the instance, and only in case it made
a prediction mistake the Perceptron update its classification rule which is a hyperplane.
Since our setting requires building a multiclass rule, we use the version described in [6]
which generalises the Perceptron to multiclass settings. We first describe the general form
of the algorithm and then discuss the modifications we performed in order to adapt it to the
sound localization problem.
To extend the Perceptron algorithm to multiclass problem we maintain hyperplanes (one
per class) denoted
87
. The algorithm works in an online fashion working on one
example at a time. On the ' th round, the algorithm gets a new instance and set the
predicted class to be the index of the hyperplane attaining the largest inner-product with
the input instance, @ D+EF GID:J
If the algorithm made a prediction error,
%
%
, it updates the set of hyperplanes.
In [6] a family of possible update
that is @
schemes was given. In this work we have used the so called uniform update which is
very simple to implement and also attained very good results. The uniform update moves
the hyperplane corresponding to the correct label 78 in the direction of and all the
hyperplanes whose inner-products were larger than 78 away from . Formally, let
" 9 " X % 7 8 We update the hyperplanes as follows,
"
(8)
"
= 8 =
%
%
C
then we keep
and if "
% intact. This update of the hyperplanes is performed
only on rounds on which there was a prediction error. Furthermore, on such rounds only
a subset of the vectors is updated and thus the algorithm is called ultraconservative. The
multiclass Perceptron algorithm is guaranteed to converge to a perfect classification rule
if the data can be classified perfectly by an unknown set of hyperplanes. When the data
cannot be classified perfectly then an alternative competitive analysis can be applied.
The problem with above algorithm is that it allows only linear classification rules. However, linear classifiers may not suffice to obtain in many applications, including the sound
localization application. We therefore incorporate kernels into the multiclass Perceptron.
A kernel is an inner-product operator B
where is the instance space (for
instance, PHAT waveforms). An explicit way to describe is via a mapping B
,' /
, ' / . Common kernels
from to an inner-products space such that , '
' /
,
/ , / .
are RBF kernels and polynomial kernels which take the form
Any learning algorithm that is based on inner-products with a weighted sum of vectors can
be converted to a kernel-based version by explicitly keeping the weighted combination of
vectors. In the case of the multiclass Perceptron we replace the update from Eq. 8 with a
?kernelized? version,
, /
"
(9)
"
= 8 = , /
%
%
Since we cannot compute , / explicitly we instead perform bookkeeping of the weights
associated with each , / and compute a inner-products using the kernel functions. For
, / with a new instance is
instance, the inner-product of a vector 3
, /
, /
3
3
,
/.
Algorithm
PHAT + Poly Kernels, D=5
PHAT + Fisher
PHAT + Peak-finding
Equalized CrossCor + Peak-finding
Err
Table 1: Summary of results of sound localization methods for a single segment.
In our experiments we found that polynomial kernel of degree yielded the best results.
The results are summarised in Table 1. We defer the discussion of the results to Sec. 5.
4 Multi-segment Classification
The accuracy of a single segment classifier is too low to make our approach practical. However, if the source of sound does not move for a period of time, we can accumulate evidence
from multiple segments in order to increase the accuracy. Due to the lack of space we only
outline the multi-segment classification procedure for the Fisher discriminant and compare
it to smoothing and averaging techniques used in the signal processing community.
In multi-segment classification we are given waveforms for which we assume that
6
. Each
the source angle did not change in this period, i.e., , - 1
78
6
small window was processed independently to give a feature vector . We then converted the waveform feature vector into a probability estimate for each discrete angle
A 6
direction, , 9
%& / using the Fisher discriminant. We next assumed that the probability estimates for consecutive windows are independent. This is of course a false assumption. However, we found that methods which compensate for the dependencies did
not yield substantial improvements. The probability density function of the entire winA
A @ , 6 9
%& /
and the ML estimation for @ is
" 6! *
dow is therefore @ , ! 9
%& / #
! $ GID:J&%'(*) A @ , ! 9
%& / We compared
@
the Maximum Likelihood decision un
der the independence assumption with the following commonly used signal processing
technique. We averaged the power spectrum and cross power spectrum of the different
windows and only then we proceeded to compute the generalized cross correlation wave +.- $ 0
form, ! " ( ,
is the average over the measurements
/ &
where +
6
3 6! * 1
in the same window, + 1
The averaged weight function for the
!
PHAT waveform is now ( ,.-0/ 129 +2- ,:-</ $ ,:-</ / 9
When using averaged power
spectrum it is also possible to define a smoothed coherent transform (SCOT) [1]. The
weight vector in this case is identical to the PHAT weight in the single segment case,
( ,.-0/ 1243 + - ,:-0/ $ ,:-0/ / + - ,:-0/ $ ,:-0/ / . Finally, we applied the classification
techniques for the single segments on the resulting (smoothed or averaged) waveform.
5 Experimental Results
In this section we report and discuss results of experiments that we performed with the
various learning algorithms for single-segments and multiple segments. Measurements
where made using the Sony ERS-210 AIBO robot. The sampling frequency was fixed to
- . )1 0 " 2
1 and the robot?s uni-directional microphone without automatic level control
was used. The robot was laid on a concrete floor in a regular office room, the room reverberations was 50687:9 0<4 . A loudspeaker, playing speech data from multiple speakers,
was placed 1 ; in front of the robot and ; above its plane, the background noise
<>9@? ACB . A PC connected through a wireless link to the robot directed its head
was =
relative to the speaker. The location of the sound source was limited to be in front of the
head (
) at a fixed constant elevation and in jumps of 1 . Therefore,
78
the number of classes, , for training is 1 . An illustration of the system is given in Fig. 2.
Algorithm
Max. Likl. PHAT + Fisher
SCOT + Fisher
Smoothed PHAT + Fisher
Smoothed PHAT + Peak-finding
SCOT + Peak-finding
Err
Table 2: Summary of results of sound localization methods for multiple segments.
Further technical details can be obtained from http://udi.benreuven.com. (MATLAB is a
trademark of Mathworks, Inc. and AIBO is a trademark of Sony and its affiliates.) For each
head direction ? segments of data were collected. Each segment is )1 0;84 long. The
segments were collected with a partial overlap of 1 ; 4 . For each direction, the measurements were divided into equal amounts of train and test measurements. The total number of
segments per class, , is . Therefore, altogether there were
%
%
segments for training and the same amount for evaluation. An FFT of size 1 was used to
generate un-normalized cross-correlations, equalized cross-correlations, and PHAT waveforms. From the transformed waveforms 1 1 samples where taken (
in Eq. 6). Extrema
locations in histograms were found using 9 9 ? 1 bins.
We used two evaluation measures for comparing the different algorithms. The first, denoted
+ !3! , is the empirical classification error that counts the number of times the predicted
(discretized) angle was different than the true angle, that is, + !3! 3 +( *
@ .
The second evaluation measure, denoted , is the average absolute( difference
between
3 (
9 @ 9 . It should be kept
the predicted angle and the true angle,
*
( same direction set as the training
in mind that the test data was obtained from the
data. Therefore,
is an appropriate evaluation measure of the errors in our experi
mental setting. However, alternative evaluation methods should be devised for general
recordings when the test signal is not confined to a finite set of possible directions.
The accuracy results with respect to both measures on the test data for the various representations and algorithms are summarized in Table 1.
It is clear from the results that traditional methods which search for extrema in the waveforms
are inferior to the discriminative methods. As a
by-product we confirmed that equalized crosscorrelations is inferior to PHAT modelling for
high SNR with strong reverberations, similar results were reported in [11]. The two discrimiFigure 2: Acquisition system overview.
native methods achieve about the same results.
Using the Perceptron algorithm with degree
achieves the best results but the difference between the Perceptron and the multiclass Fisher
discriminant is not statistically significant. It is worth noting again that we also tested linear regression algorithms. Their performance turns to be inferior to the discriminative
multiclass approaches. A possible explanation is that the multiclass methods employ multiple hyperplanes and project each class onto a different hyperplane while linear regression
methods seek a single hyperplane onto which example are projected.
Although Fisher?s discriminant and the Perceptron algorithm exhibit practically the same
performance, they have different merits. While Fisher?s discriminant is very simple
to implement and is space efficient the Perceptron is capable to adapt quickly and
achieves high accuracy even with small amounts of training data. In Fig 3 we compare the error rates of Fisher?s discriminant and the Perceptron on subsets of the training data. The Perceptron clearly outperforms Fisher?s discriminant when the number of training examples is less than but once about examples are pro
vided the two algorithms are indistinguishable. This suggests that online algorithms
may be more suitable when the sound source is stationary only for short periods.
Last we compared multi-segment results. Multisegment classification was performed by taking
? 1 consecutive measurements over a window of 0;4 during which the source location remained fix. In Table 2 we report classification results for the various multi-segment
techniques. (Since the Perceptron algorithm
used a very large number of kernels we did not
implement a multi-segment classification using
the Perceptron. We are currently conducting research on space-efficient kernel-based methods
for multi-segment classification.) Here again,
the best performing method is Fisher?s discrimFigure 3: Error rates of Fisher?s disinant that combines the scores directly without
criminant and the Perceptron for variaveraging and smoothing leads the pack. The reous training sizes.
sulting prediction accuracy of Fisher?s discriminant is good enough to make the solution practical so long as the sound source is fixed and the recording conditions do not change.
48
Perceptron
Fisher
47
46
Error Rate
45
44
43
42
41
40
39
1000
2000
3000
4000
5000
6000
Number of Examples
6 Discussion
We have demonstrated that by using discriminative methods highly accurate sound localization is achievable on a small commercial robot equipped with a binaural hearing that
are placed inside a manikin head. We have confirmed that PHAT is superior to plain crosscorrelation. For classification using multiple segments classifying the entire PHAT waveform gave better results than various techniques that smooth the power spectrum over the
segments. Our current research is focused on efficient discriminative methods for sound
localization in changing environments.
References
[1] C. H. Knapp and G. C. Carter. The generalized correlation method for estimation of time delay.
IEEE Transactions on ASSP, 24(4):320-327,1976.
[2] M. Omologo and P. Svaizer. Acoustic event localization using a crosspowerspectrum phase based
technique. Proceedings of ICASSP1994, Adelaide, Australia, 1994.
[3] T. Gustafsson and B.D. Rao. Source Localization in Reverberant Environments: Statistical Analysis. Submitted to IEEE Trans. on Speech and Audio Processing, 2000.
[4] N. Strobel and R. Rabenstein. Classification of Time Delay Estimates for Robust Speaker Localization ICASSP, Phoenix, USA, March 1999.
[5] J. Benesty Adaptive eigenvalue decomposition algorithm for passive acoustic source localization
J. Acoust. Soc. Am. 107 (1), January 2000
[6] K. Crammer and Y. Singer. Ultraconservative online algorithms for multiclass problems. In Proc.
of the 14th Annual Conf. on Computational Learning Theory, 2001.
[7] R. O. Duda, P. E. Hart. Pattern Classification. Wiley, 1973.
[8] B. Porat. A course in Digital Signal Processing. Wiley, 1997.
[9] F. Rosenblatt. The Perceptron: A probabilistic model for information storage and organization
in the brain. Psychological Review, 65:386?407, 1958.
[10] B. Widrow and M. E. Hoff. Adaptive switching circuits. 1960 IRE WESCON Convention
Record, pages 96?104, 1960.
[11] P. Aarabi, A. Mahdavi. The Relation Between Speech Segment Selectivity and Time-Delay
Estimation Accuracy. In Proc. of IEEE Conf. on Acoustics Speech and Signal Processing, 2002.
| 2151 |@word proceeded:1 version:4 achievable:1 polynomial:2 duda:1 azimuthal:1 propagate:1 seek:1 covariance:2 decomposition:1 reduction:1 score:2 denoting:1 outperforms:1 err:2 current:1 com:2 comparing:1 written:1 remove:1 treating:2 localise:1 designed:1 update:8 stationary:4 plane:1 short:5 record:1 mental:1 ire:1 quantized:1 location:18 hyperplanes:7 along:1 udi:2 gustafsson:1 combine:1 inside:1 indeed:1 expected:1 multi:7 antipodal:1 discretized:1 brain:1 inspired:2 globally:1 equipped:2 considering:2 window:6 inappropriate:1 project:1 suffice:1 circuit:1 medium:3 israel:1 gid:2 acoust:1 finding:7 extremum:5 classifier:2 control:1 engineering:1 local:1 mistake:1 switching:1 despite:1 path:1 interpolation:1 approximately:2 chose:1 suggests:1 co:1 ease:1 limited:3 statistically:1 averaged:4 directed:1 practical:2 practice:1 implement:5 procedure:1 empirical:2 significantly:1 regular:1 get:2 cannot:2 onto:2 operator:3 storage:1 demonstrated:1 jerusalem:1 independently:1 focused:1 resolution:1 simplicity:1 rule:5 array:7 variation:1 updated:1 commercial:2 us:1 element:1 native:1 bottom:1 observed:1 connected:1 substantial:1 environment:4 motivate:1 segment:35 localization:16 binaural:2 icassp:1 various:5 derivation:1 train:1 describe:9 effective:1 equalized:7 whose:2 lag:4 larger:1 valued:1 otherwise:1 transform:11 online:6 advantage:1 eigenvalue:1 quantizing:1 maximal:4 product:9 inserting:1 aligned:1 combining:1 iff:1 translate:1 achieve:2 degenerate:1 intuitive:1 categorization:1 perfect:1 ben:1 object:1 tdoa:3 widrow:2 ac:1 school:1 received:6 eq:9 strong:1 soc:1 implemented:1 c:1 predicted:3 come:1 convention:1 direction:8 waveform:27 correct:1 australia:1 bin:3 suffices:1 generalization:1 fix:1 elevation:1 biological:1 hold:1 practically:1 considered:1 normal:2 mapping:2 achieves:2 consecutive:2 estimation:4 proc:2 label:2 currently:1 largest:1 weighted:2 clearly:2 rather:2 casting:1 office:1 derived:1 focus:2 improvement:1 modelling:1 likelihood:4 am:1 entire:6 typically:1 kernelized:1 relation:1 expand:1 transformed:1 dual:1 classification:23 denoted:6 smoothing:2 hoff:2 phat:24 equal:2 once:1 sampling:2 identical:2 broad:1 peaked:1 contaminated:1 report:3 spline:1 employ:4 composed:3 individual:1 phase:4 maintain:1 organization:1 highly:1 multiply:1 evaluation:5 extreme:1 pc:1 accurate:2 capable:1 partial:1 divide:1 sulting:1 psychological:1 instance:14 classify:2 rao:1 localizing:1 hearing:1 subset:4 snr:1 uniform:2 delay:8 azimuth:8 too:1 front:2 reuven:1 reported:1 dependency:1 acb:1 combined:1 density:3 peak:16 huji:1 sequel:1 probabilistic:1 quickly:1 concrete:1 again:2 choose:1 conf:2 crosscorrelation:2 mahdavi:1 account:1 converted:3 attaining:2 sec:8 summarized:1 inc:1 explicitly:2 performed:4 break:1 closed:1 analyze:1 portion:1 wave:3 competitive:1 defer:1 contribution:1 il:1 accuracy:7 variance:1 conducting:1 efficiently:1 yield:1 directional:1 modelled:2 mere:1 confirmed:2 worth:1 classified:3 submitted:1 explain:1 nonetheless:1 acquisition:1 frequency:9 npo:1 associated:1 dimensionality:2 organized:1 shaping:1 adaptability:1 attained:1 supervised:3 response:5 improved:1 furthermore:1 stage:2 correlation:22 working:1 receives:1 dow:1 overlapping:1 lack:1 usa:1 effect:3 building:1 normalized:2 true:3 round:3 indistinguishable:1 during:1 inferior:5 speaker:3 generalized:7 outline:2 demonstrate:1 passive:1 omologo:1 pro:1 common:5 superior:1 bookkeeping:1 physical:7 overview:1 neurologically:1 phoenix:1 extend:1 accumulate:1 significant:2 measurement:10 automatic:1 had:1 moving:2 robot:9 access:1 multivariate:1 recent:1 selectivity:1 binary:1 der:1 floor:2 converge:1 paradigm:1 period:3 affiliate:1 signal:28 multiple:7 sound:21 exceeds:1 smooth:2 generalises:1 adapt:2 technical:1 cross:21 compensate:2 long:2 divided:2 devised:1 hart:1 bigger:1 prediction:9 regression:4 histogram:6 represent:1 kernel:13 achieved:1 confined:1 addition:3 background:1 interval:2 source:24 eliminates:1 pass:3 recording:2 near:2 presence:1 ter:1 noting:1 reverberant:1 easy:1 enough:1 fft:1 affect:1 independence:1 gave:2 perfectly:2 inner:7 multiclass:17 locating:1 speech:4 matlab:1 detailed:1 clear:1 amount:3 obstruction:1 processed:2 carter:1 diameter:1 crosscorrelations:1 generate:2 http:1 per:2 rosenblatt:1 summarised:1 discrete:3 changing:1 kept:1 padded:1 sum:1 angle:13 letter:2 inverse:2 vided:1 throughout:1 family:1 laid:1 decision:2 guaranteed:1 yielded:2 annual:1 fourier:6 speed:1 performing:2 loudspeaker:1 combination:2 march:1 conjugate:1 mally:1 modification:1 ener:1 taken:2 benesty:1 discus:3 count:1 turn:1 mathworks:1 singer:3 mind:2 merit:2 sony:3 end:1 multiplied:1 apply:1 away:1 spectral:3 appropriate:1 alternative:2 altogether:1 top:1 denotes:1 entertainment:2 yx:1 yoram:1 build:3 especially:1 move:3 traditional:2 exhibit:1 distance:1 separate:1 link:1 extent:2 aibo:3 discriminant:12 collected:2 assuming:2 index:2 illustration:1 hebrew:1 setup:3 difficult:1 reverberation:5 unknown:4 perform:5 upper:1 vertical:1 convolution:1 summarising:1 finite:1 supporting:1 january:1 aarabi:1 assp:1 head:12 smoothed:4 community:1 overcoming:1 pair:1 acoustic:3 coherent:1 learned:1 trans:1 bar:1 usually:2 below:1 pattern:1 including:1 max:1 explanation:1 power:5 overlap:1 suitable:1 difficulty:1 event:1 scheme:3 improve:1 brief:1 imply:1 deviate:1 prior:1 geometric:1 knapp:1 review:1 relative:5 age:1 digital:1 degree:2 experi:1 mercer:1 playing:1 classifying:1 prone:1 course:3 summary:2 placed:3 wireless:1 transpose:1 keeping:1 last:1 perceptron:22 taking:2 absolute:1 plain:2 evaluating:1 commonly:5 made:3 regressors:1 jump:1 projected:1 adaptive:2 cope:1 transaction:1 compact:1 uni:1 keep:1 ml:3 wescon:1 generalising:1 assumed:2 conclude:1 discriminative:8 neurology:1 spectrum:11 search:3 continuous:1 ultraconservative:2 un:2 porat:1 table:5 channel:4 pack:1 robust:1 investigated:1 complex:2 necessarily:1 poly:1 ehud:1 domain:5 did:3 noise:8 arrival:2 body:2 fig:3 fashion:2 wiley:2 fails:1 position:1 hanning:1 explicit:1 weighting:3 scot:3 formula:2 remained:1 trademark:2 er:1 evidence:1 incorporating:1 quantization:1 false:1 simply:1 corresponds:2 goal:2 likl:1 rbf:1 room:3 labelled:1 fisher:19 replace:1 change:5 hyperplane:5 averaging:1 microphone:19 called:4 conservative:1 total:1 experimental:4 intact:1 formally:2 internal:1 crammer:1 adelaide:1 incorporate:1 evaluate:2 audio:1 tested:2 |
1,265 | 2,152 | Interpreting Neural Response Variability as
Monte Carlo Sampling of the Posterior
Patrik O. Hoyer and Aapo Hyv?arinen
Neural Networks Research Centre
Helsinki University of Technology
P.O. Box 9800, FIN-02015 HUT, Finland
http://www.cis.hut.fi/phoyer/
[email protected]
Abstract
The responses of cortical sensory neurons are notoriously variable, with
the number of spikes evoked by identical stimuli varying significantly
from trial to trial. This variability is most often interpreted as ?noise?,
purely detrimental to the sensory system. In this paper, we propose an alternative view in which the variability is related to the uncertainty, about
world parameters, which is inherent in the sensory stimulus. Specifically, the responses of a population of neurons are interpreted as stochastic samples from the posterior distribution in a latent variable model. In
addition to giving theoretical arguments supporting such a representational scheme, we provide simulations suggesting how some aspects of
response variability might be understood in this framework.
1 Introduction
During the past half century, a wealth of data has been collected on the response properties
of cortical sensory neurons. The majority of this research has focused on how the mean
firing rates of individual neurons depend on the sensory stimulus. Similarly, mathematical
models have mainly focused on describing how the mean firing rate could be computed
from the input. One aspect which this research does not address is the high variability of
cortical neural responses. The trial-to-trial variation in responses to identical stimuli are
significant [1, 2], and several trials are typically required to get an adequate estimate of the
mean firing rate.
The standard interpretation is that this variability reflects ?noise? which limits the accuracy
of the sensory system [2, 3]. In the standard model, the firing rate is given by
where
rate
stimulus
noise
(1)
is the ?tuning function? of the cell in question. Here, the magnitude of the noise
may depend on the stimulus. Experimental results [1, 2] seem to suggest that the amount of
variability depends only on the mean firing rate, i.e. stimulus , and not on the particular
Current address: 4 Washington Place, Rm 809, New York, NY 10003, USA
stimulus that evoked it. Specifically, spike count variances tend to grow in proportion to
spike count means [1, 2]. This has been taken as evidence for something like a Poisson
process for neural firing.
This standard view is not completely satisfactory. First, the exquisite sensitivity and the
reliability of many peripheral neurons (see, e.g. [3]) show that neurons in themselves need
not be very unreliable. In vitro experiments [4] also suggest that the large variability does
not have its origin in the neurons themselves, but is a property of intact cortical circuits.
One is thus tempted to point at synaptic ?background? activity as the culprit, attributing the
variability of individual neurons to variable inputs. This seems reasonable, but it is not
quite clear why such modulation of firing should be considered meaningless noise rather
than reflecting complex neural computations.
Second, the above model does a poor job of explaining neural responses in the phenomenon
known as ?visual competition?: When viewing ambiguous (bistable) figures, perception,
and the responses of many neurons with it, oscillates between two distinct states (for a
review, see [5]). In other words, a single stimulus can yield two very different firing rates
in a single neuron depending on how the stimulus is interpreted. In the above model, this
means that either (a) the noise term needs to have a bimodal distribution, or (b) we are
forced to accept the fact that neurons can be tuned to stimulus interpretations, rather than
stimuli themselves. The former solution is clearly unattractive. The latter seems sensible,
but we have then simply transformed the problem of oscillating firing rates into a problem
of oscillating interpretations: Why should there be variability (over time, and over trials) in
the interpretation of a stimulus?
What would be highly desirable is a theoretical framework in which the variability of responses could be shown to have a specific purpose. One suggestion [6] is that variability
could improve the signal to noise ratio through a phenomenon known as ?stochastic resonance?. Another recent suggestion is that variability contributes to the contrast invariance
of visual neurons [7].
In this paper, we will propose an alternative explanation for the variability of neural responses. This hypothesis attempts to account for both aspects of variability described
above: the Poisson-like ?noise? and the oscillatory responses to ambiguous stimuli. Our
suggestion is based on the idea that cortical circuits implement Bayesian inference in latent variable models [8, 9, 10]. Specifically, we propose that neural firing rates might be
viewed as representing Monte Carlo samples from the posterior distribution over the latent
variables, given the observed input. In this view, the response variability is related to the
uncertainty, about world parameters, which is inherent in any stimulus. This representation would allow not only the coding of parameter values but also of their uncertainties.
The latter could be accomplished by pooling responses over time, or over a population of
redundant cells.
Our proposal has a direct connection to Monte Carlo methods widely used in engineering.
These methods use built-in randomness to solve difficult problems that cannot be solved
analytically. In particular, such methods are one of the main options for performing approximate inference in Bayesian networks [11]. With that in mind, it is perhaps even a bit
surprising that Monte Carlo sampling has not, to our knowledge, previously been suggested
as an explanation for the randomness of neural responses.
Although the approach proposed is not specific to sensory modality, we will here, for concreteness, exclusively concentrate on vision. We shall start by, in the next section, reviewing the basic probabilistic approach to vision. Then we will move on to further explain the
proposal of this contribution.
2 The latent variable approach to vision
2.1 Bayesian models of high-level vision
Recently, a growing number of researchers have argued for a probabilistic approach to
vision, in which the functioning of the visual system is interpreted as performing Bayesian
inference in latent variable models, see e.g. [8, 9, 10]. The basic idea is that the visual
input is seen as the observed data in a probabilistic generative model. The goal of vision
is to estimate the latent (i.e. unobserved or hidden) variables that caused the given sensory
stimulation.
In this framework, there are a number of world parameters that contribute to the observed
data. These could be, for example, object identities, dimensions and locations, surface
properties, lighting direction, and so forth. These parameters are not directly available to
the sensory system, but must be estimated from the effects that they have on the images
projected onto the retinas. Collecting all the unknown world variables into the vector
and all sensory data into the vector , the probability that a given set of world parameters
caused a given sensory stimulus is
(2)
describes
is known
is the prior probability of the set of world parameters , and
where
how sensory data is generated from the world parameters. The distribution
as the posterior distribution.
A specific perceptual task then consists of estimating some subset of the world variables,
given the observed data [10]. In face recognition, for example, one wants to know the
identity of a person but one does not care about the specific viewpoint or the direction of
lighting. Note, however, that sometimes one might specifically want to estimate viewpoint
or lighting, disregarding identity, so one cannot just automatically throw out that information [10]. In a latent variable model, all relevant information is contained in the complete
posterior distribution identity viewpoint lighting sensory data . To estimate the identity
one must use the marginal posterior identity sensory data , obtained by integrating out
the viewpoint and lighting variables. Bayesian models of high-level vision model the visual
system as performing these types of computations, but typically do not specify how they
might be neurally implemented.
2.2 Neural network models of low-level vision
This probabilistic approach has not only been suggested as an abstract framework for vision, but in fact also as a model for interpreting actual neural firing patterns in the early
visual cortex [12, 13]. In this line of research, the hypothesis is that the activity of individual neurons can be associated with hidden state variables, and that the neural circuitry
implements probabilistic inference.1
The model of Olshausen and Field [12], known as sparse coding or independent component analysis (ICA) [14], depending on the viewpoint taken, is perhaps the most influential latent variable model of early visual processing to date. The hidden variables
are
independent and sparse, such as is given, for instance, by the double-sided exponential
distribution
. The observed data vector is then given by a
,
linear combination of the , plus additive isotropic Gaussian noise. That is,
1
Here, it must be stressed that in these low-level neural network models, the hidden variables that
the neurons represent are not what we would typically consider to be the ?causal? variables of a visual
scene. Rather, they are low-level visual features similar to the optimal stimuli of neurons in the early
visual cortex. The belief is that more complex hierarchical models will eventually change this.
where is a matrix of
model parameters (weights), and is Gaussian with zero mean and
covariance matrix
.
How does this abstract probabilistic model relate to neural processing? Olshausen and
Field showed that when the model parameters are estimated (learned) from natural image
data, the basis vectors (columns of ) come to resemble V1 simple cell receptive fields.
Moreover, the latent variables relate to the activities of the corresponding cells. Specifically, Olshausen and Field suggested [12] that the firing rates of the neurons correspond
to
the maximum a posteriori (MAP) estimate of the latent variables, given the image input:
.
An important problem with this kind of a MAP representation is that it attempts to represent a complex posterior distribution using only a single point (at the maximum). Such a
representation cannot adequately represent multimodal posterior distributions, nor does it
provide any way of coding the uncertainty of the value (the width of the peak). Many other
proposed neural representations of probabilities face similar problems [11] (however, see
[15] for a recent interesting approach to representing distributions). Indeed, it has been said
[10, 16] that how probabilities actually are represented in the brain is one of the most important unanswered questions in the probabilistic approach to perception. In the next section
we suggest an answer based on the idea that probability distributions might be represented
using response variability.
3 Neural responses as samples from the posterior distribution?
As discussed in the previous section, the distribution of primary interest to a sensory system is the posterior distribution over world parameters. In all but absolutely trivial models,
computing and representing such a distribution requires approximative methods, of which
one major option is Monte Carlo methods. These generate stochastic samples from a given
distribution, without explicitly calculating it, and such samples can then be used to approximately represent or perform computations on that distribution [11].
Could the brain use a Monte Carlo approach to perform Bayesian inference? If neural
firing rates are used (even indirectly) to represent continuous-valued latent variables, one
possibility would be for firing rate variability to represent a probability distribution over
these variables. Here, there are two main possibilities:
(a) Variability over time. A single neuron could represent a continuous distribution if
its firing rate fluctuated over time in accordance with the distribution to be represented. At each instant in time, the instantaneous firing rate would be a random
sample from the distribution to be represented.
(b) Variability over neurons. A distribution could be instantaneously represented if
the firing rate of each neuron in a pool of identical cells was independently and
randomly drawn from the distribution to be represented.
Note that these are not exclusive, both types of variability could potentially coexist. Also
note that both cases lead to trial-to-trial variability, as all samples are assumed independent.
Both possibilities have their advantages. The first option is much more efficient in terms
of the number of cells required, which is particularly important for representing highdimensional distributions. In this case, dependencies between variables can naturally be
represented as temporal correlations between neurons representing different parameters.
This is not nearly as straightforward for case (b). On the other hand, in terms of processing
speed, this latter option is clearly preferred to the former. Any decisions should optimally
be based on the whole posterior distribution, and in case (a) this would require collecting
samples over an extended period of time.
10
10
10
10
1
1
1
1
0.1
0.1
1
10
0.1
0.1
1
10
0.1
0.1
1
10
0.1
0.1
1
10
Figure 1: Variance of response versus mean response, on log-log axes, for 4 representative
model neurons. Each dot gives the mean (horizontal axis) and variance (vertical axis) of
the response of the model neuron in question to one particular stimulus. Note that the scale
of responses is completely arbitrary.
We will now explain how both aspects of response variability described in the introduction
can be understood in this framework. First, we will show how a simple mean-variance relationship can arise through sampling in the independent component analysis model. Then,
we will consider how the variability associated with the phenomenon of visual competion
can be interpreted using sampling.
3.1 Example 1: Posterior sampling in ICA
Here, we sample the posterior distribution in the ICA model of natural images, and show
how this might relate to the conspicious variance-mean relation of neural response variability. First, we used standard ICA methods [17] to estimate a complete basis for the
-pixel natural image patches. Motivated by
40-dimensional principal subspace of
the non-negativity of neural firing rates we modified the model to assume single-sided exponential priors
[18], and augmented the basis so that a pair of neurons
coded separately for the positive and negative parts of each original independent component. We then took 50 random natural image patches and sampled the posterior distributions for all 50 patches , taking a total of 1000 samples in each case. 2
From the 1000 collected samples, we calculated the mean and variance of the response of
each neuron to each stimulus separately. We then plotted the variance against the mean
independently for each neuron in log-log coordinates. Figure 1 shows the plots from 4
randomly selected neurons. The crucial thing to note is that, as for real neurons [1], the
variance of the response is systematically related to the mean response, and does not seem
to depend on the particular stimulus used to elicit a given mean response. This feature of
neural variability is perhaps the single most important reason to believe that the variability
is meaningless noise inherent in neural firing; yet we have shown that something like this
might arise through sampling in a simple probabilistic model.
Following [1, 2], we fitted lines to the plots, modeling the variance as var
mean .
Over the whole population (80 model neurons), the mean values of and were
and
, with population standard deviations
and (respectively). Although these
values do not actually match those obtained from physiology (most reports give values of
between 1 and 2, and close to 1, see [1, 2]), this is to be expected. First, the values of these
parameters probably depend on the specifics of the ICA model, such as its dimensionality
and the noise level; we did not optimize these to attempt to fit physiology. Second, and
more importantly, we do not believe that ICA is an exact model of V1 function. Rather, the
visual cortex would be expected to employ a much more complicated, hierarchical, image
2
This was accomplished using a Markov Chain Monte Carlo method, as described in the Appendix. However, the technical details of this method are not very relevant to this argument.
model. Thus, our main goal was not to show that the particular parameters of the variancemean relation could be explained in this framework, but rather the surprising fact that such
a simple relation might arise as a result of posterior sampling in a latent variable model.
3.2 Example 2: Visual competition as sampling
As described in the introduction, in addition to the mean-variance relationship observed
throughout the visual cortex, a second sort of variability is that observed in visual competition. This phenomenon arises when viewing a bistable figure, such as the famous Necker
cube or Rubin?s vase/face figure. These figures each have two interpretations (explanations) that both cannot reasonably explain the image simultaneously. In a latent variable
image model, this corresponds to the case of a bimodal posterior distribution.
When such figures are viewed, the perception oscillates between the two interpretations (for
a review of this phenomenon, see [5]). This corresponds to jumping from mode to mode
in the posterior distribution. This can directly be interpreted as sampling of the posterior.
When the stimulus is modified so that one interpretations is slightly more natural than the
other one, the former is dominant for a relatively longer period compared with the latter
(again, see [5]), just as proper sampling takes relatively more samples from the mode which
has larger probability mass. Although the above might be considered purely ?perceptual?
sampling, animal studies indicate that especially in higher-level visual areas many neurons
modulate their responses in sync with the animal?s perceptions [5, 19]. This link proves
that some form of sampling is clearly taking place on the level of neural firing rates as well.
Note that this phenomenon might be considered as evidence for sampling scheme (a) and
against (b). If we instantaneously could represent whole distributions, we should be able to
keep both interpretations in mind simultaneously. This is in fact (weak) evidence against
any scheme of representing whole distributions instantaneously, by the same logic.
4 Conclusions
One of the key unanswered questions in theoretical neuroscience seems to be: How are
probabilities represented by the brain? In this paper, we have proposed that probability distributions might be represented using response variability. If true, this would also present
a functional explanation for the significant amount of cortical neural ?noise? observed. Although it is clear that the variability degrades performance on many perceptual tasks of the
laboratory, it might well be that it plays an important function in everyday sensory tasks.
Our proposal would be one possible way in which it might do so.
Do actual neurons employ such a computational scheme? Although our arguments and
simulations suggest that it might be possible (and should be kept in mind), future research
will be needed to answer that question. As we see it, key experiments would compare
measured firing rate variability statistics (single unit variances, or perhaps two-unit covariances) to those predicted by latent variable models. Of particular interest are cases where
contextual information reduces the uncertainty inherent in a given stimulus; our hypothesis
predicts that in such cases neural variability is also reduced.
A final question concerns how neurons might actually implement Monte Carlo sampling
in practice. Because neurons cannot have global access to the activity of all other neurons
in the population, the only possibility seems to be something akin to Gibbs sampling [20].
Such a scheme might require only relatively local information and could thus conceivably
be implemented in actual neural networks.
Acknowledgements ? Thanks to Paul Hoyer, Jarmo Hurri, Bruno Olshausen, Liam Paninski, Phil Sallee, Eero Simoncelli, and Harri Valpola for discussions and comments.
References
[1] A. F. Dean. The variability of discharge of simple cells in the cat striate cortex. Experimental
Brain Research, 44:437?440, 1981.
[2] D. J. Tolhurst, J. A. Movshon, and A. F. Dean. The statistical reliability of signals in single
neurons in cat and monkey visual cortex. Vision Research, 23:775?785, 1983.
[3] A. J. Parker and W. T. Newsome. Sense and the single neuron: Probing the physiology of
perception. Annual Review of Neuroscience, 21:227?277, 1998.
[4] G. R. Holt, W. R. Softky, C. Koch, and R. J. Douglas. Comparison of discharge variability
in vitro and in vivo in cat visual cortex neurons. Journal of Neurophysiology, 75:1806?1814,
1996.
[5] R. Blake and N. K. Logothetis. Visual competition. Nature Reviews Neuroscience, 3:13?21,
2002.
[6] M. Rudolph and A. Destexhe. Do neocortical pyramidal neurons display stochastic resonance?
Journal of Computational Neuroscience, 11:19?42, 2001.
[7] J. S. Anderson, I. Lampl, D. C. Gillespie, and D. Ferster. The contribution of noise to contrast
invariance of orientation tuning in cat visual cortex. Science, 290:1968?1972, 2000.
[8] D. C. Knill and W. Richards, editors. Perception as Bayesian Inference. Cambridge University
Press, 1996.
[9] R. P. N. Rao, B. A. Olshausen, and M. S. Lewicki, editors. Probabilistic Models of the Brain.
MIT Press, 2002.
[10] D. Kersten and P. Schrater. Pattern inference theory: A probabilistic approach to vision. In
R. Mausfeld and D. Heyer, editors, Perception and the Physical World. Wiley & Sons, 2002.
[11] P. Dayan. Recognition in hierarchical models. In F. Cucker and M. Shub, editors, Foundations
of Computational Mathematics. Springer, Berlin, Germany, 1997.
[12] B. A. Olshausen and D. J. Field. Sparse coding with an overcomplete basis set: A strategy
employed by V1? Vision Research, 37:3311?3325, 1997.
[13] R. P. N. Rao and D. H. Ballard. Predictive coding in the visual cortex: a functional interpretation
of some extra-classical receptive field effects. Nature Neuroscience, 2(1):79?87, 1999.
[14] A. J. Bell and T. J. Sejnowski. The ?independent components? of natural scenes are edge filters.
Vision Research, 37:3327?3338, 1997.
[15] R. S. Zemel, P. Dayan, and A. Pouget. Probabilistic interpretation of population codes. Neural
Computation, 10(2):403?430, 1998.
[16] H. B. Barlow. Redundancy reduction revisited. Network: Computation in Neural Systems,
12:241?253, 2001.
[17] A. Hyv?arinen. Fast and robust fixed-point algorithms for independent component analysis.
IEEE Trans. on Neural Networks, 10(3):626?634, 1999.
[18] P. O. Hoyer. Modeling receptive fields with non-negative sparse coding. In E. De Schutter,
editor, Computational Neuroscience: Trends in Research 2003. Elsevier, Amsterdam, 2003. In
press.
[19] N. K. Logothetis and J. D. Schall. Neuronal correlates of subjective visual perception. Science,
245:761?763, 1989.
[20] S. Geman and D. Geman. Stochastic relaxation, gibbs distributions, and the bayesian restoration
of images. IEEE Transactions on Pattern Analysis and Machine Intelligence, 6:721?741, 1984.
Appendix: MCMC sampling of the non-negative ICA posterior
The posterior probability of , upon observing , is given by
(3)
Taking the (natural) logarithm yields
(4)
where
is a vector of all ones. The crucial thing to note is that this function is quadratic in
. Thus, the posterior distribution has the form of a gaussian, except that of course it is only
defined for non-negative . Rejection sampling might look tempting, but unfortunately does
not work well in high dimensions. Thus, we will instead opt for a Markov Chain Monte
Carlo approach. Implementing Gibbs sampling [20] is quite straightforward. The posterior
distribution of
, given and all other hidden variables , is a one-dimensional density
that we will call cut-gaussian,
if
!"+$* #&%
%(
,
In this case, we have the following parameter values:
3
0
21
!
1
if .- if /
')(
!
1
(5)
and
54
(6)
Here, 1 denotes the 6 :th column of , and
3 denotes the current state vector but with
set to zero. Sampling from such a one-dimensional distribution is relatively simple. Just as
one can sample the corresponding (uncut) gaussian by taking uniformly distributed samples
and passing them through the inverse of the gaussian cumulative
on the interval
distribution function, the same can be done for a cut-gaussian distribution by constraining
the uniform sampling interval suitably.
Hence Gibbs sampling is feasible, but, as is well known, Gibbs sampling exhibits problems
when there are significant correlations between the sampled variables. Thus we choose to
use a sampling scheme based on a rotated co-ordinate system. The basic idea is to update
the state vector not in the directions of the component axes, as in standard Gibbs sampling,
but rather in the directions of the eigenvectors of . Thus we start by calculating
these eigenvectors, and cycle through them one at a time. Denoting the current unit-length
eigenvector to be updated 7 we have as a function of the step length ,
7
const
8
7
7
7
7
(7)
Again, note how this is a quadratic function of . Again, the non-negativity constraints
on require us to sample a cut-gaussian distribution. But this time there is an additional
complication: When the basis is overcomplete, some of the eigenvectors will be associated
with zero eigenvalues, and the logarithmic probability will be linear instead of quadratic.
Thus, in such a case we must sample a cut-exponential distribution,
:9
0=<
if
if
if
;
>/;
-
(8)
Like in the cut-gaussian case, this can be done by uniformly sampling the corresponding
interval and then applying the inverse of the exponential cumulative distribution function.
In summary: We start by calculating the eigensystem of the matrix , and set the
state vector to random non-negative values. Then we cycle through the eigenvectors
indefinitely, sampling from cut-gaussian or cut-exponential distributions depending on
the eigenvalue corresponding to the current eigenvector 7 , and updating the state vector
to
7 . MATLAB code performing and verifying this sampling is available at:
http://www.cis.hut.fi/phoyer/code/samplingpack.tar.gz
| 2152 |@word neurophysiology:1 trial:8 proportion:1 seems:4 suitably:1 hyv:2 simulation:2 covariance:2 reduction:1 exclusively:1 tuned:1 denoting:1 past:1 subjective:1 current:4 contextual:1 surprising:2 culprit:1 yet:1 must:4 additive:1 plot:2 update:1 half:1 generative:1 selected:1 intelligence:1 isotropic:1 indefinitely:1 tolhurst:1 contribute:1 location:1 revisited:1 complication:1 mathematical:1 direct:1 consists:1 sync:1 indeed:1 expected:2 ica:7 themselves:3 nor:1 growing:1 brain:5 automatically:1 actual:3 estimating:1 moreover:1 circuit:2 mass:1 what:2 kind:1 interpreted:6 monkey:1 eigenvector:2 unobserved:1 temporal:1 collecting:2 oscillates:2 rm:1 unit:3 positive:1 understood:2 engineering:1 accordance:1 local:1 limit:1 uncut:1 firing:21 modulation:1 approximately:1 might:17 plus:1 evoked:2 co:1 liam:1 practice:1 implement:3 area:1 elicit:1 bell:1 significantly:1 physiology:3 word:1 integrating:1 holt:1 suggest:4 get:1 cannot:5 onto:1 coexist:1 close:1 applying:1 kersten:1 www:2 optimize:1 map:2 dean:2 phil:1 straightforward:2 independently:2 focused:2 pouget:1 importantly:1 population:6 century:1 unanswered:2 variation:1 coordinate:1 discharge:2 updated:1 play:1 logothetis:2 exact:1 approximative:1 hypothesis:3 origin:1 trend:1 recognition:2 particularly:1 updating:1 richards:1 cut:7 predicts:1 geman:2 observed:8 solved:1 verifying:1 cycle:2 depend:4 reviewing:1 predictive:1 purely:2 upon:1 completely:2 basis:5 multimodal:1 represented:9 cat:4 harri:1 distinct:1 forced:1 fast:1 monte:9 sejnowski:1 zemel:1 quite:2 widely:1 solve:1 valued:1 larger:1 statistic:1 rudolph:1 final:1 advantage:1 eigenvalue:2 took:1 propose:3 relevant:2 date:1 representational:1 forth:1 competition:4 everyday:1 double:1 oscillating:2 rotated:1 object:1 depending:3 measured:1 job:1 throw:1 implemented:2 predicted:1 resemble:1 come:1 indicate:1 concentrate:1 direction:4 filter:1 stochastic:5 viewing:2 bistable:2 implementing:1 argued:1 arinen:2 require:3 opt:1 hut:4 considered:3 koch:1 blake:1 circuitry:1 major:1 finland:1 early:3 purpose:1 reflects:1 instantaneously:3 mit:1 clearly:3 gaussian:10 modified:2 rather:6 tar:1 varying:1 ax:2 mainly:1 contrast:2 sense:1 posteriori:1 inference:7 elsevier:1 dayan:2 typically:3 accept:1 hidden:5 relation:3 transformed:1 germany:1 pixel:1 orientation:1 resonance:2 animal:2 marginal:1 field:7 cube:1 washington:1 sampling:27 identical:3 look:1 nearly:1 future:1 report:1 stimulus:22 inherent:4 employ:2 retina:1 randomly:2 simultaneously:2 individual:3 attempt:3 interest:2 highly:1 possibility:4 chain:2 edge:1 jumping:1 logarithm:1 plotted:1 causal:1 overcomplete:2 theoretical:3 fitted:1 instance:1 column:2 modeling:2 rao:2 newsome:1 restoration:1 deviation:1 subset:1 sallee:1 uniform:1 optimally:1 dependency:1 answer:2 person:1 thanks:1 peak:1 sensitivity:1 density:1 probabilistic:11 pool:1 cucker:1 again:3 choose:1 suggesting:1 account:1 de:1 coding:6 caused:2 explicitly:1 depends:1 view:3 observing:1 start:3 sort:1 option:4 complicated:1 vivo:1 contribution:2 accuracy:1 variance:11 yield:2 correspond:1 necker:1 weak:1 bayesian:8 famous:1 carlo:9 notoriously:1 researcher:1 lighting:5 randomness:2 oscillatory:1 explain:3 synaptic:1 against:3 naturally:1 associated:3 sampled:2 knowledge:1 dimensionality:1 actually:3 reflecting:1 higher:1 jarmo:1 response:29 specify:1 done:2 box:1 anderson:1 just:3 correlation:2 hand:1 horizontal:1 mode:3 perhaps:4 believe:2 olshausen:6 usa:1 effect:2 true:1 barlow:1 functioning:1 adequately:1 analytically:1 former:3 hence:1 laboratory:1 satisfactory:1 during:1 width:1 ambiguous:2 eigensystem:1 complete:2 neocortical:1 interpreting:2 image:10 instantaneous:1 fi:3 recently:1 stimulation:1 functional:2 vitro:2 physical:1 discussed:1 interpretation:10 schrater:1 significant:3 cambridge:1 gibbs:6 tuning:2 mathematics:1 similarly:1 centre:1 bruno:1 reliability:2 dot:1 access:1 cortex:9 surface:1 longer:1 something:3 dominant:1 posterior:22 recent:2 showed:1 accomplished:2 seen:1 additional:1 care:1 employed:1 exquisite:1 redundant:1 period:2 signal:2 tempting:1 neurally:1 desirable:1 simoncelli:1 reduces:1 technical:1 match:1 coded:1 aapo:1 basic:3 vision:13 poisson:2 sometimes:1 represent:8 bimodal:2 cell:7 proposal:3 addition:2 background:1 want:2 separately:2 interval:3 wealth:1 grow:1 pyramidal:1 modality:1 crucial:2 extra:1 meaningless:2 probably:1 comment:1 pooling:1 tend:1 thing:2 seem:2 call:1 constraining:1 destexhe:1 fit:1 idea:4 motivated:1 akin:1 movshon:1 york:1 passing:1 adequate:1 matlab:1 clear:2 eigenvectors:4 amount:2 reduced:1 http:2 generate:1 estimated:2 neuroscience:6 shall:1 key:2 redundancy:1 drawn:1 douglas:1 kept:1 v1:3 relaxation:1 concreteness:1 inverse:2 uncertainty:5 place:2 throughout:1 phoyer:2 reasonable:1 patch:3 decision:1 appendix:2 bit:1 display:1 quadratic:3 annual:1 activity:4 constraint:1 helsinki:1 scene:2 aspect:4 speed:1 argument:3 performing:4 relatively:4 influential:1 peripheral:1 combination:1 poor:1 describes:1 slightly:1 son:1 conceivably:1 schall:1 explained:1 sided:2 taken:2 previously:1 describing:1 count:2 eventually:1 needed:1 mind:3 know:1 available:2 hierarchical:3 indirectly:1 alternative:2 original:1 lampl:1 denotes:2 instant:1 const:1 calculating:3 giving:1 especially:1 prof:1 classical:1 move:1 question:6 spike:3 receptive:3 primary:1 exclusive:1 degrades:1 striate:1 strategy:1 said:1 hoyer:4 exhibit:1 detrimental:1 subspace:1 softky:1 link:1 valpola:1 berlin:1 majority:1 sensible:1 collected:2 trivial:1 reason:1 code:3 length:2 relationship:2 ratio:1 difficult:1 unfortunately:1 potentially:1 relate:3 negative:5 shub:1 proper:1 unknown:1 perform:2 vertical:1 neuron:37 markov:2 fin:1 supporting:1 extended:1 variability:34 arbitrary:1 ordinate:1 pair:1 required:2 connection:1 learned:1 trans:1 address:2 able:1 suggested:3 perception:8 pattern:3 built:1 explanation:4 belief:1 gillespie:1 natural:7 vase:1 representing:6 scheme:6 improve:1 technology:1 axis:2 negativity:2 gz:1 patrik:2 review:4 prior:2 acknowledgement:1 suggestion:3 interesting:1 versus:1 var:1 foundation:1 rubin:1 viewpoint:5 editor:5 systematically:1 course:1 summary:1 allow:1 explaining:1 face:3 taking:4 sparse:4 distributed:1 dimension:2 cortical:6 world:10 calculated:1 cumulative:2 sensory:16 projected:1 correlate:1 transaction:1 approximate:1 schutter:1 preferred:1 unreliable:1 keep:1 logic:1 global:1 assumed:1 hurri:1 eero:1 continuous:2 latent:14 why:2 nature:2 reasonably:1 ballard:1 robust:1 contributes:1 complex:3 did:1 main:3 whole:4 noise:13 arise:3 paul:1 knill:1 augmented:1 neuronal:1 representative:1 parker:1 ny:1 probing:1 wiley:1 exponential:5 perceptual:3 specific:5 disregarding:1 evidence:3 unattractive:1 concern:1 ci:2 magnitude:1 conspicious:1 rejection:1 attributing:1 logarithmic:1 simply:1 paninski:1 visual:22 amsterdam:1 contained:1 lewicki:1 springer:1 corresponds:2 fluctuated:1 modulate:1 viewed:2 goal:2 identity:6 ferster:1 tempted:1 feasible:1 change:1 specifically:5 except:1 uniformly:2 principal:1 total:1 invariance:2 experimental:2 intact:1 highdimensional:1 latter:4 stressed:1 arises:1 absolutely:1 mcmc:1 phenomenon:6 |
1,266 | 2,153 | Prediction and Semantic Association
Thomas L. Griffiths & Mark Steyvers
Department of Psychology
Stanford University, Stanford, CA 94305-2130
{gruffydd,msteyver}@psych.stanford.edu
Abstract
We explore the consequences of viewing semantic association as
the result of attempting to predict the concepts likely to arise in a
particular context. We argue that the success of existing accounts
of semantic representation comes as a result of indirectly addressing
this problem, and show that a closer correspondence to human data
can be obtained by taking a probabilistic approach that explicitly
models the generative structure of language.
1
Introduction
Many cognitive capacities, such as memory and categorization, can be analyzed as
systems for efficiently predicting aspects of an organism's environment [1]. Previously, such analyses have been concerned with memory for facts or the properties
of objects, where the prediction task involves identifying when those facts might
be needed again, or what properties novel objects might possess. However, one of
the most challenging tasks people face is linguistic communication. Engaging in
conversation or reading a passage of text requires retrieval of a variety of concepts
from memory in response to a stream of information. This retrieval task can be
facilitated by predicting which concepts are likely to be needed from their context,
having efficiently abstracted and stored the cues that support these predictions.
In this paper, we examine how understanding the problem of predicting words
from their context can provide insight into human semantic association, exploring
the hypothesis that the association between words is at least partially affected
by their statistical relationships. Several researchers have argued that semantic
association can be captured using high-dimensional spatial representations , with
the most prominent such approach being Latent Semantic Analysis (LSA) [5]. We
will describe this procedure, which indirectly addresses the prediction problem. We
will then suggest an alternative approach which explicitly models the way language
is generated and show that this approach provides a better account of human word
association data than LSA, although the two approaches are closely related. The
great promise of this approach is that it illustrates how we might begin to relax some
of the strong assumptions about language made by many corpus-based methods.
We will provide an example of this, showing results from a generative model that
incorporates both sequential and contextual information.
2
Latent Semantic Analysis
Latent Semantic Analysis addresses the prediction problem by capturing similarity
in word usage: seeing a word suggests that we should expect to see other words
with similar usage patterns. Given a corpus containing W words and D documents,
the input to LSA is a W x D word-document co-occurrence matrix F in which fwd
corresponds to the frequency with which word w occurred in document d. This
matrix is transformed to a matrix G via some function involving the term frequency
fwd and its frequency across documents fw .. Many applications of LSA in cognitive
science use the transformation
gwd = IOg{fwd
+ 1}(1 -
Hw)
H
- _
2:D_
Wlog{W}
d-l f w
f w.
logD
w -
'
(1)
where Hw is the normalized entropy of the distribution over documents for each
word. Singular value decomposition (SVD) is applied to G to extract a lower
dimensional linear subspace that captures much of the variation in usage across
words. The output of LSA is a vector for each word, locating it in the derived
subspace. The association between two words is typically assessed using the cosine of
the angle between their vectors, a measure that appears to produce psychologically
accurate results on a variety of tasks [5] . For the tests presented in this paper,
we ran LSA on a subset of the TASA corpus, which contains excerpts from texts
encountered by children between first grade and the first year of college. Our subset
used all D = 37651 documents, and the W = 26414 words that occurred at least
ten times in the whole corpus, with stop words removed. From this we extracted a
500 dimensional representation, which we will use throughout the paper. 1
3
The topic model
Latent Semantic Analysis gives results that seem consistent with human judgments
and extracts information relevant to predicting words from their contexts, although
it was not explicitly designed with prediction in mind. This relationship suggests
that a closer correspondence to human data might be obtained by directly attempting to solve the prediction task. In this section, we outline an alternative approach
that involves learning a probabilistic model of the way language is generated. One
generative model that has been used to outperform LSA on information retrieval
tasks views documents as being composed of sets of topics [2,4]. If we assume that
the words that occur in different documents are drawn from T topics, where each
topic is a probability distribution over words, then we can model the distribution
over words in anyone document as a mixture of those topics
T
P(Wi)
= LP(Wil zi =j)P(Zi
=j)
(2)
j=l
where Zi is a latent variable indicating the topic from which the ith word was drawn
and P(wilzi = j) is the probability of the ith word under the jth topic. The words
likely to be used in a new context can be determined by estimating the distribution
over topics for that context, corresponding to P(Zi).
Intuitively, P(wlz = j) indicates which words are important to a topic, while P(z)
is the prevalence of those topics within a document. For example, imagine a world
where the only topics of conversation are love and research. We could then express
IThe dimensionality of the representation is an important parameter for both models in
this paper. LSA performed best on the word association task with around 500 dimensions,
so we used the same dimensionality for the topic model.
the probability distribution over words with two topics, one relating to love and the
other to research. The content of the topics would be reflected in P(wlz = j): the
love topic would give high probability to words like JOY, PLEASURE, or HEART, while
the research topic would give high probability to words like SCIENCE, MATHEMATICS,
or EXPERIMENT. Whether a particular conversation concerns love, research, or
the love of research would depend upon its distribution over topics, P(z), which
determines how these topics are mixed together in forming documents.
Having defined a generative model, learning topics becomes a statistical problem.
The data consist of words w = {Wl' ... , w n }, where each Wi belongs to some document di , as in a word-document co-occurrence matrix. For each document we
have a multinomial distribution over the T topics, with parameters ()(d), so for a
word in document d, P(Zi = j) = ();d;). The jth topic is represented by a multinomial distribution over the W words in the vocabulary, with parameters 1/i), so
P(wilzi = j) = 1>W. To make predictions about new documents, we need to assume
a prior distribution on the parameters (). Existing parameter estimation algorithms
make different assumptions about (), with varying results [2,4]. Here, we present a
novel approach to inference in this model, using Markov chain Monte Carlo with a
symmetric Dirichlet(a) prior on ()(di) for all documents and a symmetric Dirichlet(,B)
prior on 1>(j) for all topics. In this approach we do not need to explicitly represent
the model parameters: we can integrate out () and 1>, defining the model simply in
terms of the assignments of words to topics indicated by the Zi'
Markov chain Monte Carlo is a procedure for obtaining samples from complicated
probability distributions, allowing a Markov chain to converge to the taq~et distribution and then drawing samples from the states of that chain (see [3]). We
use Gibbs sampling, where each state is an assignment of values to the variables
being sampled, and the next state is reached by sequentially sampling all variables
from their distribution when conditioned on the current values of all other variables
and the data. We will sample only the assignments of words to topics, Zi. The
conditional posterior distribution for Zi is given by
'1
)
P( Zi=)Zi ,wex
where
Z- i
is the assignment of all
Zk
+ (3 n(di) + a
-',}
-',}
(d ' )
n_i,j + W (3 n_i,. + Ta
n eW;)
(.)
such that k
f:.
(3)
i, and n~~:j is the number
of words assigned to topic j that are the same as w, n~L is the total number of
words assigned to topic j, n~J,j is the number of words from document d assigned
to topic j, and n~J. is the total number of words in document d, all not counting
the assignment of the current word Wi. a,,B are free parameters that determine how
heavily these distributions are smoothed.
We applied this algorithm to our subset of the TASA corpus, which contains n =
5628867 word tokens. Setting a = 0.1,,B = 0.01 we obtained 100 samples of 500
topics, with 10 samples from each of 10 runs with a burn-in of 1000 iterations and
a lag of 100 iterations between samples. 2 Each sample consists of an assignment of
every word token to a topic, giving a value to each Zi. A subset of the 500 topics
found in a single sample are shown in Table 1. For each sample we can compute
2Random numbers were generated with the Mersenne Twister, which has an extremely
deep period [6]. For each run, the initial state of the Markov chain was found using an
on-line version of Equation 3.
FEEL
FEELINGS
FEELING
ANGRY
WAY
THINK
SHOW
FEELS
PEOPLE
FRIENDS
THINGS
MIGHT
HELP
HAPPY
FELT
LOVE
ANGER
BEING
WAYS
FEAR
MUSIC
BALL
GAME
TEAM
PLAY
DANCE
PLAYS
STAGE
PLAYED
BAND
AUDIENCE
MUSICAL
DANCING
RHYTHM
PLAYING
THEATER
DRUM
ACTORS
SHOW
BALLET
ACTOR
DRAMA
SONG
PLAY
BASEBALL
FOOTBALL
PLAYERS
GAMES
PLAYING
FIELD
PLAYED
PLAYER
COACH
BASKETBALL
SPORTS
HIT
BAT
TENNIS
TEAMS
SOCCER
SCIENCE
STUDY
SCIENTISTS
SCIENTIFIC
KNOWLEDGE
WORK
CHEMISTRY
RESEARCH
BIOLOGY
MATHEMATICS
LABORATORY
STUDYING
SCIENTIST
PHYSICS
FIELD
STUDIES
UNDERSTAND
STUDIED
SCIENCES
MANY
WORKERS
WORK
LABOR
JOBS
WORKING
WORKER
WAGES
FACTORY
JOB
WAGE
SKILLED
PAID
CONDITIONS
PAY
FORCE
MANY
HOURS
EMPLOYMENT
EMPLOYED
EMPLOYERS
FORCE
FORCES
MOT IO N
BODY
GRAVITY
MASS
PULL
NEWTON
OBJECT
LAW
DIRECTION
MOVING
REST
FALL
ACTING
MOMENTUM
DISTANCE
GRAVITATIONAL
PUSH
VELOCITY
Table 1: Each column shows the 20 most probable words in one of the 500 topics
obtained from a single sample. The organization of the columns and use of boldface
displays the way in which polysemy is captured by the model.
the posterior predictive distribution (and posterior mean for q/j)) :
P(wl z
4
= j, z, w) =
J
P(wl z
(.) ( Iz,
= j, ? J
0)
)P(? J
w) d?
( 0)
J
+ (3
+ W (3
n (W)
= _(;=,.J)_ _
nj
(4)
Predicting word association
We used both LSA and the topic model to predict the association between pairs
of words, comparing these results with human word association norms collected by
Nelson, McEvoy and Schreiber [7]. These word association norms were established
by presenting a large number of participants with a cue word and asking them to
name an associated word in response. A total of 4544 of the words in these norms
appear in the set of 26414 taken from the TASA corpus.
4.1
Latent Semantic Analysis
In LSA, the association between two words is usually measured using the cosine
of the angle between their vectors. We ordered the associates of each word in the
norms by their frequencies , making the first associate the word most commonly
given as a response to the cue. For example, the first associate of NEURON is BRAIN.
We evaluated the cosine between each word and the other 4543 words in the norms ,
and then computed the rank of the cosine of each of the first ten associates, or
all of the associates for words with less than ten. The results are shown in Figure
1. Small ranks indicate better performance, with a rank of one meaning that the
target word had the highest cosine. The median rank of the first associate was 32,
and LSA correctly predicted the first associate for 507 of the 4544 words.
4.2
The topic model
The probabilistic nature of the topic model makes it easy to predict the words likely
to occur in a particular context. If we have seen word WI in a document, then we
can determine the probability that word W2 occurs in that document by computing
P( w2IwI). The generative model allows documents to contain multiple topics, which
450
400
1_
LSA - cosine
LSA - inner product
Topi c model
D
1
350
300
II
250
200
150
l;r
100
50
o
lin
2
3
4
5
6
7
8
9
10
Associate number
Figure 1: Performance of different methods of prediction on the word association
task. Error bars show one standard error, estimated with 1000 bootstrap samples.
is extremely important to capturing the complexity of large collections of words
and computing the probability of complete documents. However, when comparing
individual words it is more effective to assume that they both come from a single
topic. This assumption gives us
(5)
z
where we use Equation 4 for P(wlz) and P(z) is uniform, consistent with the symmetric prior on e, and the subscript in Pi (w2lwd indicates the restriction to a single
topic. This estimate can be computed for each sample separately, and an overall
estimate obtained by averaging over samples. We computed Pi (w2Iwi) for the 4544
words in the norms, and then assessed the rank of the associates in the resulting
distribution using the same procedure as for LSA. The results are shown in Figure
1. The median rank for the first associate was 32, with 585 of the 4544 first associates exactly correct. The probabilistic model performed better than LSA, with
the improved performance becoming more apparent for the later associates .
4.3
Discussion
The central problem in modeling semantic association is capturing the interaction
between word frequency and similarity of word usage. Word frequency is an important factor in a variety of cognitive tasks, and one reason for its importance is its
predictive utility. A higher observed frequency means that a word should be predicted to occur more often. However, this effect of frequency should be tempered by
the relationship between a word and its semantic context . The success of the topic
model is a consequence of naturally combining frequency information with semantic
similarity: when a word is very diagnostic of a small number of topics, semantic
context is used in prediction. Otherwise, word frequency plays a larger role.
The effect of word frequency in the topic model can be seen in the rank-order
correlation of the predicted ranks of the first associates with the ranks predicted
by word frequency alone , which is p = 0.49. In contrast, the cosine is used in LSA
because it explicitly removes the effect of word frequency, with the corresponding
correlation being p = -0.01. The cosine is purely a measure of semantic similarity,
which is useful in situations where word frequency is misleading, such as in tests of
English fluency or other linguistic tasks, but not necessarily consistent with human
performance. This measure was based in the origins of LSA in information retrieval ,
but other measures that do incorporate word frequency have been used for modeling
psychological data. We consider one such measure in the next section.
5
Relating LSA and the topic model
The decomposition of a word-document co-occurrence matrix provided by the topic
model can be written in a matrix form similar to that of LSA. Given a worddocument co-occurrence matrix F, we can convert the columns into empirical estimates of the distribution over words in each document by dividing each column
by its sum. Calling this matrix P, the topic model approximates it with the nonnegative matrix factorization P ~ ?O, where column j of ? gives 4/j) , and column d
of 0 gives ()(d). The inner product matrix ppT is proportional to the empirical estimate of the joint distribution over words P(WI' W2)' We can write ppT ~ ?OOT ?T,
corresponding to P(WI ,W2) = L z"Z 2 P(wIl zdP(W2Iz2)P(ZI,Z2) , with OOT an empirical estimate of P(ZI , Z2)' The theoretical distribution for P(ZI, Z2) is proportional to 1+ 0::, where I is the identity matrix, so OOT should be close to diagonal.
The single topic assumption removes the off-diagonal elements, replacing OOT with
I to give PI (Wl ' W2) ex: ??T.
By comparison, LSA transforms F to a matrix G via Equation 1, then the SVD
gives G ~ UDV T for some low-rank diagonal D. The locations of the words along
the extracted dimensions are X = UD. If the column sums do not vary extensively,
the empirical estimate of the joint distribution over words specified by the entries in
G will be approximately P(WI,W2) ex: GG T . The properties of the SVD guarantee
that XX T , the matrix of inner products among the word vectors , is the best lowrank approximation to GG T in terms of squared error. The transformations in
Equation 1 are intended to reduce the effects of word frequency in the resulting
representation, making XX T more similar to ??T.
We used the inner product between word vectors to predict the word association
norms, exactly as for the cosine. The results are shown in Figure 1. The inner
product initially shows worse performance than the cosine, with a median rank
of 34 for the first associate and 500 exactly correct, but performs better for later
associates. The rank-order correlation with the predictions of word frequency for
the first associate was p = 0.46, similar to that for the topic model. The rankorder correlation between the ranks given by the inner product and the topic model
was p = 0.81, while the cosine and the topic model correlate at p = 0.69. The
inner product and PI (w2lwd in the topic model seem to give quite similar results,
despite being obtained by very different procedures. This similarity is emphasized
by choosing to assess the models with separate ranks for each cue word, since this
measure does not discriminate between joint and conditional probabilities. While
the inner product is related to the joint probability of WI and W2, PI (w2lwd is a
conditional probability and thus allows reasonable comparisons of the probability
of W2 across choices of WI , as well as having properties like asymmetry that are
exhibited by word association.
"syntax"
HE
YOU
THEY
I
SHE
WE
IT
PEOPLE
EVERYONE
OTHERS
SCIENTISTS
SOMEONE
WHO
NOBODY
ONE
SOMETHING
ANYONE
EVERYBODY
SOME
THEN
ON
AT
INTO
FROM
WITH
THROUGH
OVER
AROUND
AGAINST
ACROSS
UPON
TOWARD
UNDER
ALONG
NEAR
BEHIND
OFF
ABOVE
DOWN
BEFORE
BE
MAKE
GET
HAVE
GO
TAKE
DO
FIND
USE
SEE
HELP
KEEP
GIVE
LOOK
COME
WORK
MOVE
LIVE
EAT
BECOME
"semantics"
SAID
ASKED
THOUGHT
TOLD
SAYS
MEANS
CALLED
CRIED
S HOWS
ANSWERED
TELLS
REPLIED
SHOUTED
EXPLAINED
LAUGHED
MEANT
WROTE
SHOWED
BELIEVED
WHISPERED
MAP
NORTH
EARTH
SOUTH
POLE
MAPS
EQUATOR
WEST
LINES
EAST
AUSTRALIA
GLOBE
POLES
HEMISPHERE
LATITUDE
PLACES
LAND
WORLD
COMPASS
CONTINE NTS
DOCTOR
PATIENT
HEALTH
HOSPITAL
MEDICAL
CARE
PATIENTS
NURSE
DOCTORS
MEDICINE
NURSING
TREATMENT
NURSES
PHYSICIAN
HOSPITALS
DR
S ICK
ASSISTANT
EMERGENCY
PRACTICE
Table 2: Each column shows the 20 most probable words in one of the 48 "syntactic"
states of the hidden Markov model (four columns on the left) or one of the 150
"semantic" topics (two columns on the right) obtained from a single sample.
6
Exploring more complex generative models
The topic model, which explicitly addresses the problem of predicting words from
their contexts, seems to show a closer correspondence to human word association
than LSA. A major consequence of this analysis is the possibility that we may be
able to gain insight into some of the associative aspects of human semantic memory
by exploring statistical solutions to this prediction problem. In particular, it may
be possible to develop more sophisticated generative models of language that can
capture some of the important linguistic distinctions that influence our processing
of words. The close relationship between LSA and the topic model makes the latter
a good starting point for an exploration of semantic association, but perhaps the
greatest potential of the statistical approach is that it illustrates how we might go
about relaxing some of the strong assumptions made by both of these models.
One such assumption is the treatment of a document as a "bag of words" , in which
sequential information is irrelevant. Semantic information is likely to influence only
a small subset of the words used in a particular context, with the majority of the
words playing functional syntactic roles that are consistet across contexts. Syntax is
just as important as semantics for predicting words, and may be an effective means
of deciding if a word is context-dependent. In a preliminary exploration of the
consequences of combining syntax and semantics in a generative model for language,
we applied a simple model combining the syntactic structure of a hidden Markov
model (HMM) with the semantic structure of the topic model. Specifically, we used
a third-order HMM with 50 states in which one state marked the start or end of
a sentence, 48 states each emitted words from a different multinomial distribution,
and one state emitted words from a document-dependent multinomial distribution
corresponding to the topic model with T = 150. We estimated parameters for this
model using Gibbs sampling, integrating out the parameters for both the HMM and
the topic model and sampling a state and a topic for each of the 11821091 word
tokens in the corpus. 3 Some of the state and topic distributions from a single sample
after 1000 iterations are shown in Table 2. The states of the HMM accurately picked
out many of the functional classes of English syntax, while the state corresponding
to the topic model was used to capture the context-specific distributions over nouns.
3This larger number is a result of including low frequency and stop words.
Combining the topic model with the HMM seems to have advantages for both: no
function words are absorbed into the topics, and the HMM does not need to deal
with the context-specific variation in nouns. The model also seems to do a good job
of generating topic-specific text - we can clamp the distribution over topics to pick
out those of interest, and then use the model to generate phrases. For example, we
can generate phrases on the topics of research ( "the chief wicked selection of research
in the big months" , "astronomy peered upon your scientist's door", or "anatomy
established with principles expected in biology") , language ("he expressly wanted
that better vowel"), and the law ("but the crime had been severely polite and
confused" , or "custody on enforcement rights is plentiful"). While these phrases
are somewhat nonsensical , they are certainly topical.
7
Conclusion
Viewing memory and categorization as systems involved in the efficient prediction
of an organism's environment can provide insight into these cognitive capacities.
Likewise, it is possible to learn about human semantic association by considering
the problem of predicting words from their contexts. Latent Semantic Analysis
addresses this problem, and provides a good account of human semantic association.
Here, we have shown that a closer correspondence to human data can be obtained
by taking a probabilistic approach that explicitly models the generative structure
of language, consistent with the hypothesis that the association between words
reflects their probabilistic relationships. The great promise of this approach is the
potential to explore how more sophisticated statistical models of language, such as
those incorporating both syntax and semantics, might help us understand cognition.
Acknowledgments
This work was generously supported by the NTT Communications Sciences Laboratories.
We used Mersenne Twister code written by Shawn Cokus, and are grateful to Touchstone
Applied Science Associates for making available the TASA corpus, and to Josh Tenenbaum
for extensive discussions on this topic.
References
[1] J. R. Anderson. The Adaptive Character of Thought. Erlbaum, Hillsdale, NJ, 1990.
[2] D . M. Blei, A. Y. Ng, and M. 1. Jordan . Latent Dirichlet allocation. In T. G. Dietterich,
S. Becker, and Z. Ghahramani, eds, Advances in Neural Information Processing Systems
14, 2002.
[3] W . R. Gilks, S. Richardson, and D. J . Spiegelhalter, eds. Markov Chain Monte Carlo
in Practice. Chapman and Hall, Suffolk, 1996.
[4] T . Hofmann. Probabilistic Latent Semantic Indexing. In Proceedings of the TwentySecond Annual International SIGIR Conference, 1999.
[5] T. K. Landauer and S. T. Dumais. A solution to Plato's problem: The Latent Semantic
Analysis theory of acquisition, induction, and representation of knowledge. Psychological
Review, 104:211- 240, 1997.
[6] M. Matsumoto and T . Nishimura. Mersenne twister: A 623-dimensionally equidistributed uniform pseudorandom number generator. ACM Transactions on Modeling and
Computer Simulation, 8:3- 30, 1998.
[7] D. L. Nelson , C. L. McEvoy, and T. A. Schreiber. The University of South Florida
word association norms. http://www. usf. edu/FreeAssociation, 1999.
| 2153 |@word version:1 norm:8 seems:3 nonsensical:1 simulation:1 decomposition:2 paid:1 pick:1 initial:1 plentiful:1 contains:2 document:27 existing:2 current:2 contextual:1 comparing:2 z2:3 nt:1 written:2 hofmann:1 wanted:1 remove:2 designed:1 joy:1 alone:1 generative:9 cue:4 mcevoy:2 ith:2 blei:1 provides:2 location:1 along:2 skilled:1 become:1 consists:1 expected:1 udv:1 examine:1 love:6 grade:1 brain:1 touchstone:1 considering:1 becomes:1 begin:1 estimating:1 provided:1 xx:2 confused:1 mass:1 what:1 psych:1 astronomy:1 transformation:2 nj:2 guarantee:1 every:1 gravity:1 exactly:3 hit:1 medical:1 lsa:22 appear:1 before:1 scientist:4 consequence:4 io:1 despite:1 severely:1 subscript:1 becoming:1 approximately:1 might:7 burn:1 studied:1 suggests:2 challenging:1 someone:1 co:4 relaxing:1 factorization:1 bat:1 acknowledgment:1 gilks:1 drama:1 practice:2 prevalence:1 bootstrap:1 procedure:4 empirical:4 oot:4 thought:2 word:104 integrating:1 griffith:1 seeing:1 suggest:1 get:1 close:2 selection:1 context:16 live:1 influence:2 restriction:1 www:1 map:2 go:2 starting:1 sigir:1 twentysecond:1 identifying:1 insight:3 theater:1 pull:1 steyvers:1 mersenne:3 variation:2 feel:2 imagine:1 play:4 heavily:1 target:1 hypothesis:2 origin:1 engaging:1 associate:17 velocity:1 element:1 observed:1 role:2 capture:3 removed:1 highest:1 ran:1 environment:2 wil:2 complexity:1 asked:1 employment:1 depend:1 grateful:1 ithe:1 predictive:2 purely:1 upon:3 baseball:1 replied:1 joint:4 represented:1 describe:1 effective:2 monte:3 tell:1 choosing:1 apparent:1 lag:1 stanford:3 solve:1 larger:2 say:1 relax:1 drawing:1 football:1 otherwise:1 richardson:1 think:1 syntactic:3 associative:1 advantage:1 clamp:1 interaction:1 product:8 relevant:1 combining:4 asymmetry:1 produce:1 categorization:2 generating:1 object:3 help:3 friend:1 develop:1 fluency:1 measured:1 lowrank:1 job:3 strong:2 dividing:1 predicted:4 involves:2 come:3 indicate:1 direction:1 anatomy:1 closely:1 correct:2 twister:3 exploration:2 human:12 australia:1 viewing:2 hillsdale:1 argued:1 fwd:3 d_:1 preliminary:1 probable:2 exploring:3 gravitational:1 around:2 hall:1 deciding:1 great:2 cognition:1 predict:4 major:1 vary:1 earth:1 estimation:1 assistant:1 bag:1 wl:4 schreiber:2 reflects:1 ick:1 generously:1 varying:1 linguistic:3 derived:1 she:1 rank:14 indicates:2 contrast:1 inference:1 dependent:2 typically:1 initially:1 hidden:2 transformed:1 semantics:4 overall:1 among:1 spatial:1 noun:2 field:2 having:3 ng:1 sampling:4 chapman:1 biology:2 look:1 suffolk:1 anger:1 others:1 ppt:2 composed:1 individual:1 intended:1 vowel:1 organization:1 interest:1 possibility:1 certainly:1 analyzed:1 mixture:1 behind:1 chain:6 accurate:1 closer:4 worker:2 theoretical:1 psychological:2 column:10 modeling:3 asking:1 compass:1 assignment:6 phrase:3 pole:2 addressing:1 subset:5 entry:1 uniform:2 erlbaum:1 stored:1 mot:1 dumais:1 international:1 probabilistic:7 physic:1 off:2 told:1 physician:1 together:1 again:1 central:1 squared:1 containing:1 dr:1 worse:1 cognitive:4 account:3 potential:2 chemistry:1 north:1 explicitly:7 stream:1 performed:2 view:1 later:2 picked:1 reached:1 doctor:2 start:1 participant:1 complicated:1 ass:1 freeassociation:1 musical:1 who:1 efficiently:2 likewise:1 judgment:1 accurately:1 carlo:3 researcher:1 ed:2 against:1 acquisition:1 frequency:18 involved:1 naturally:1 associated:1 di:3 nurse:2 stop:2 sampled:1 gain:1 treatment:2 conversation:3 knowledge:2 dimensionality:2 sophisticated:2 appears:1 ta:1 higher:1 reflected:1 response:3 improved:1 evaluated:1 anderson:1 just:1 stage:1 correlation:4 working:1 replacing:1 indicated:1 perhaps:1 scientific:1 usage:4 dietterich:1 name:1 concept:3 normalized:1 wlz:3 contain:1 effect:4 assigned:3 symmetric:3 laboratory:2 laughed:1 semantic:25 deal:1 game:2 basketball:1 everybody:1 rhythm:1 soccer:1 cosine:11 prominent:1 gg:2 syntax:5 presenting:1 outline:1 complete:1 performs:1 passage:1 logd:1 meaning:1 novel:2 multinomial:4 functional:2 equidistributed:1 association:23 organism:2 occurred:2 relating:2 approximates:1 he:2 gibbs:2 mathematics:2 language:9 had:2 nobody:1 moving:1 actor:2 similarity:5 tennis:1 something:1 posterior:3 showed:1 belongs:1 hemisphere:1 irrelevant:1 dimensionally:1 success:2 tempered:1 captured:2 seen:2 care:1 somewhat:1 employed:1 converge:1 determine:2 period:1 ud:1 ii:1 multiple:1 ntt:1 believed:1 retrieval:4 lin:1 iog:1 prediction:13 involving:1 globe:1 patient:2 psychologically:1 represent:1 iteration:3 equator:1 audience:1 separately:1 singular:1 median:3 w2:7 rest:1 posse:1 exhibited:1 south:2 plato:1 thing:1 incorporates:1 seem:2 jordan:1 emitted:2 near:1 counting:1 door:1 coach:1 concerned:1 easy:1 variety:3 quite:1 psychology:1 zi:14 drum:1 inner:8 reduce:1 shawn:1 whether:1 utility:1 becker:1 song:1 locating:1 deep:1 useful:1 transforms:1 ten:3 band:1 extensively:1 tenenbaum:1 generate:2 http:1 outperform:1 estimated:2 diagnostic:1 correctly:1 write:1 promise:2 iz:1 affected:1 express:1 dancing:1 four:1 drawn:2 year:1 convert:1 sum:2 run:2 facilitated:1 angle:2 you:1 employer:1 throughout:1 reasonable:1 place:1 excerpt:1 ballet:1 capturing:3 pay:1 emergency:1 angry:1 played:2 correspondence:4 display:1 encountered:1 nonnegative:1 annual:1 occur:3 your:1 calling:1 felt:1 aspect:2 answered:1 anyone:2 extremely:2 gruffydd:1 attempting:2 eat:1 pseudorandom:1 department:1 ball:1 across:5 character:1 wi:9 lp:1 making:3 intuitively:1 explained:1 indexing:1 heart:1 taken:1 equation:4 previously:1 needed:2 mind:1 enforcement:1 end:1 studying:1 available:1 indirectly:2 occurrence:4 alternative:2 florida:1 expressly:1 thomas:1 dirichlet:3 newton:1 music:1 medicine:1 giving:1 ghahramani:1 move:1 occurs:1 diagonal:3 said:1 subspace:2 distance:1 pleasure:1 separate:1 capacity:2 majority:1 hmm:6 nelson:2 topic:63 argue:1 collected:1 reason:1 boldface:1 toward:1 induction:1 code:1 relationship:5 happy:1 allowing:1 neuron:1 markov:7 matsumoto:1 n_i:2 defining:1 situation:1 communication:2 team:2 topical:1 smoothed:1 pair:1 specified:1 extensive:1 sentence:1 crime:1 distinction:1 established:2 hour:1 address:4 able:1 bar:1 usually:1 pattern:1 latitude:1 reading:1 including:1 memory:5 everyone:1 greatest:1 force:3 predicting:8 misleading:1 spiegelhalter:1 extract:2 health:1 text:3 prior:4 understanding:1 review:1 law:2 expect:1 mixed:1 proportional:2 allocation:1 generator:1 integrate:1 wage:2 consistent:4 principle:1 playing:3 pi:5 land:1 token:3 supported:1 free:1 english:2 jth:2 understand:2 fall:1 taking:2 face:1 dimension:2 vocabulary:1 world:2 usf:1 msteyver:1 made:2 commonly:1 collection:1 adaptive:1 feeling:2 correlate:1 transaction:1 keep:1 wrote:1 abstracted:1 sequentially:1 corpus:8 landauer:1 latent:10 chief:1 table:4 nature:1 zk:1 learn:1 ca:1 obtaining:1 necessarily:1 complex:1 polysemy:1 whole:1 big:1 arise:1 child:1 body:1 west:1 wlog:1 momentum:1 factory:1 third:1 hw:2 down:1 specific:3 emphasized:1 showing:1 concern:1 consist:1 incorporating:1 sequential:2 importance:1 illustrates:2 tasa:4 conditioned:1 push:1 entropy:1 simply:1 explore:2 likely:5 forming:1 absorbed:1 josh:1 labor:1 ordered:1 nursing:1 partially:1 sport:1 fear:1 corresponds:1 determines:1 extracted:2 acm:1 conditional:3 identity:1 marked:1 month:1 content:1 fw:1 determined:1 specifically:1 acting:1 averaging:1 total:3 called:1 discriminate:1 hospital:2 svd:3 player:2 east:1 ew:1 indicating:1 college:1 mark:1 people:3 support:1 latter:1 assessed:2 meant:1 incorporate:1 whispered:1 dance:1 ex:2 |
1,267 | 2,154 | Margin-Based Algorithms
for Information Filtering
Nicol`o Cesa-Bianchi
DTI, University of Milan
via Bramante 65
26013 Crema, Italy
[email protected]
Alex Conconi
DTI, University of Milan
via Bramante 65
26013 Crema, Italy
[email protected]
Claudio Gentile
CRII, Universit`a dell?Insubria
Via Ravasi, 2
21100 Varese, Italy
[email protected]
Abstract
In this work, we study an information filtering model where the relevance
labels associated to a sequence of feature vectors are realizations of an
unknown probabilistic linear function. Building on the analysis of a restricted version of our model, we derive a general filtering rule based on
the margin of a ridge regression estimator. While our rule may observe
the label of a vector only by classfying the vector as relevant, experiments
on a real-world document filtering problem show that the performance
of our rule is close to that of the on-line classifier which is allowed to
observe all labels. These empirical results are complemented by a theoretical analysis where we consider a randomized variant of our rule and
prove that its expected number of mistakes is never much larger than that
of the optimal filtering rule which knows the hidden linear model.
1 Introduction
Systems able to filter out unwanted pieces of information are of crucial importance for
several applications. Consider a stream of discrete data that are individually labelled as
?relevant? or ?nonrelevant? according to some fixed relevance criterion; for instance, news
about a certain topic, emails that are not spam, or fraud cases from logged data of user
behavior. In all of these cases, a filter can be used to drop uninteresting parts of the stream,
forwarding to the user only those data which are likely to fulfil the relevance criterion.
From this point of view, the filter may be viewed as a simple on-line binary classifier.
However, unlike standard on-line pattern classification tasks, where the classifier observes
the correct label after each prediction, here the relevance of a data element is known only if
the filter decides to forward that data element to the user. This learning protocol with partial feedback is known as adaptive filtering in the Information Retrieval community (see,
e.g., [14]). We formalize the filtering problem as follows. Each element of an arbitrary
data sequence is characterized by a feature vector
and an associated relevance
label (say,
for relevant and
for nonrelevant). At each time
,
the filtering system observes the -th feature vector
and must decide whether or not to
forward it. If the data is forwarded, then its relevance label is revealed to the system,
The research was supported by the European Commission under the KerMIT Project No. IST2001-25431.
which may use this information to adapt the filtering criterion. If the data is not forwarded,
its relevance label remains hidden. We call
the -th instance of the data sequence and
the -th example. For simplicity, we assume
the pair
for all . There
are two kinds of errors the filtering system can make in judging the relevance of a feature
vector . We say that an example
and is classified
is a false positive if
as relevant by the system; similarly, we say that
is a false negative if
and
is classified as nonrelevant by the system. Although false negatives remain unknown,
the filtering system is scored according to the overall number of wrong relevance judgements it makes. That is, both false positives and false negatives are counted as mistakes.
In this paper, we study the filtering problem under the assumption that relevance judgements are generated using an unknown probabilistic linear function. We design filtering
rules that maintain a linear hypothesis and use the margin information to decide whether
to forward the next instance. Our performance measure is the regret; i.e., the number of
wrong judgements made by a filtering rule over and above those made by the rule knowing
the probabilistic function used to generate judgements. We show finite-time (nonasymptotical) bounds on the regret that hold for arbitrary sequences of instances. The only other
results of this kind we are aware of are those proven in [9] for the apple tasting model.
Since in the apple tasting model relevance judgements are chosen adversarially rather than
probabilistically, we cannot compare their bounds with ours. We report some preliminary
experimental results which might suggest the superiority of our methods as opposed to the
general transformations developed within the apple tasting framework. As a matter of fact,
these general transformations do not take margin information into account.
In Section 2, we introduce our probabilistic relevance model and make some preliminary
observations. In Section 3, we consider a restricted version of the model within which
we prove a regret bound for a simple filtering rule called SIMPLE - FIL. In Section 4, we
generalize this filtering rule and show its good performance on the Reuters Corpus Volume
1. The algorithm employed, which we call RIDGE - FIL, is a linear least squares algorithm
inspired by [2]. In that section we also prove, within the unrestricted probabilistic model,
a regret bound for the randomized variant P - RIDGE - FIL of the general filtering rule. Both
RIDGE - FIL and its randomized variant can be run with kernels [13] and adapted to the case
when the unknown linear function drifts with time.
2 Learning model, notational conventions and preliminaries
The relevance of
is given by a
(where
-valued random variable
means ?relevant?) such that there exists a fixed and unknown vector
,
,
. Hence
for which
for all
is relevant with probability
. The random variables
are assumed to be independent, whereas we do not make any assumption on the way the sequence !
! is
generated. In this model, we want to perform almost as well as the algorithm that knows
and forwards
if and only if "# . We consider linear-threshold filtering algorithms that predict the value of
through SGN $&%
is a
(' *) , where %
dynamically updated weight vector which might be intended as the current approximation
to , and ' is a suitable time-changing ?confidence? threshold. For any fixed sequence
of instances, we use +
to denote the margin
and ,+
to
!
!
denote the margin %
. We define the expected regret of the linear-threshold filtering algorithm at time as -.&
,+ /' 012 3-.4
+ 015 . We observe that in the conditional
-76 -probability space where ,+
8' is given we have
- 6 4
8'
;- 6 4
,+
;- 6 4
+
,+
('
092 8- 6 &
+
0:2 (- 6 4
92 8- 6 4
+
+
0:5
012<5 ,+
015?@ ,+
A'
8'
B+
=
+
C
=
ED +
>
D5 ,+
8'
+
=
>
where we use to denote the Bernoulli random variable which is 1 if and only if predicate is true. Integrating over all possible values of ,+ 8' we obtain
-.&
('
,+
015 (-4
+
015
D+
=
D+
D-. ,+
D-.BD ,+
,+
=
=
8'
+
8'
A+
('
5
D@
+
(1)
D+
D
(2)
where the last inequality is Markov?s. These (in)equalities will be used in Sections 3 and 4.2
for the analysis of SIMPLE - FIL and P - RIDGE - FIL algorithms.
3 A simplified model
We start by analyzing a restricted model where each data element has the same unknown
probability of being relevant and we want to perform almost as well as the filtering rule
that consistently does the optimal action (i.e., always forwards if :
and never forwards otherwise). The analysis of this model is used in Section 4 to guide the design of
good filtering rules for the unrestricted model.
and let ,
be the sample average of
, where
is
Let
,
the number of forwarded data elements in the first time steps and , is the fraction of
true positives among the elements that have been forwarded. Obviously, the optimal
rule forwards if and only if
. Consider instead the empirical rule that forwards if
and only if ,
. This rule makes a mistake only when
!,
= . To make
the probability of this event go to zero with , it suffices that -BD ,
CD = D
CD
as , which can only happen if increases quickly enough with . Hence, data
should be forwarded (irrespective to the sign of the estimate ,
) also when the confidence
level for ,
gets too small with respect to . A problem in this argument is that large
deviation bounds require
for making -.D ,
CD D
CD small. But in
our case
is unknown. To fix this, we use the condition
,
. This looks
dangerous, as we use the empirical value of ,
to control the large deviations of ,
itself; however, we will show that this approach indeed works. An algorithm, which we
call SIMPLE - FIL, implementing the above line of reasoning takes the form of the following
'
, where '
"! # %$ . The
simple rule: forward if and only if ,
expected regret at time of SIMPLE - FIL is defined as the probability that SIMPLE - FIL makes
a mistake at time minus the probability that the optimal filtering rule makes a mistake,
that is -4
,
'
0 2
-.&
0 5 . The next result shows a logarithmic bound
on this regret.
of time
Theorem 1 The expected cumulative regret of SIMPLE - FIL after any number (
steps is at most & 2CD
CD '(! ) .
Proof sketch. We can bound the actual regret after
the definition of the filtering rule we have
time steps as follows. From (1) and
*
*
.-
,
,+
,+
, +
21 *,+ 435 % $
6"!
87 *, + 43
,
9 % $
6(!
875:
;6< =;>
*
-.&
4,
8'
/0:5
D CD
-
-
=
,
4,
8'
D CD
D CD
-4
=
015
A
-
!,
=
,
=
0/
Without loss of generality, assume
. We now bound =; < and 6
; >
, we have that ;6<
Since %$ = "! ,
implies that
%$
for some . Hence we can write
separately.
implies
% $
, +* 3 %$
6(!
9 %$ 7 ,+* * + 3 6
"! 7
* *% $
(! -
9/
, + +
* *%$ -
/ for
"!
,+ % $+
* * -
/
, + +
Applying Chernoff-Hoeffding [11] bounds to
, which is a sum of -valued
independent random variables, we obtain
=;<
,+ % $+
!
We now bound ;6> by adapting a technique from [8]. Let "#%$ '& "()$
*
+/.
& * + , )$ $ *- +. We have
*
*
, ,
;>
-
"
$
/
,+
,+ 10 " $ $ 3 2
*,+ 3 , ,
$ 9 % $
6(!
87
*
* % $
*
, + +54 -
" )$
/ , + 3 ,
$ %$
6(!
7
;76 ;78
Applying Chernoff-Hoeffding bounds again, we get =;96
, + (!
Finally, one can easily verify that ;:8
. Piecing everything together we get the desired
;
result.
; <
=
=
,
=
=
,
=
=
D =
D,
D CD
=
D,
D =
D CD
=
D,
CD>
D CD
,
=
!
-
2D ,
CD>
=
D CD
@
,
,
=
D,
CD>
D CD
,
,
,
=
D,
CD@
D CD
.
,
=
,
=
C
4 Linear least squares filtering
In order to generalize SIMPLE - FIL to the original (unrestricted) learning model described
in Section 2, we need a low-variance estimate of the target vector . Let
be the matrix
whose columns are the forwarded feature vectors after the first time steps and let
be the vector of corresponding observed relevance labels (the index will be momentarily
dropped). Note that
holds. Consider the least squares estimator
of , where
is the pseudo-inverse of
. For all belonging to the column
E
space of , this is an unbiased estimator of , that is B
/
To remove the assumption on , we make
full rank by adding
the identity . This also allows us to replace the pseudo-inverse with the standard inverse,
<
@<A< B%<A<
K
>= ?
<
E<A< B
<D<
<
GF @<D< !BC<H=JI
<D<
=
@<A< !BC<D=
E<A< B%< >=
RIDGE-FULL
RIDGE-FIL
FREQUENCY x 10
1
F-MEASURE
0.8
0.6
0.4
0.2
0
34 CATEGORIES
-measure for each one of the 34 filtering tasks. The -measure is defined by
, where is precision (fraction of relevant documents among the forwarded ones) and is recall
(fraction of forwarded documents among the relevant ones). In the plot, the filtering rule RIDGE - FIL
is compared with RIDGE - FULL which sees the correct label after each classification. While precision
and recall of RIDGE - FULL are balanced, RIDGE - FIL?s recall is higher than precision due to the need
of forwarding more documents than believed relevant. This in order to make the confidence of the
estimator converge to 1 fast enough. Note that, in some cases, this imbalance causes RIDGE - FIL to
achieve a slightly better -measure than RIDGE - FULL.
Figure 1:
@K
<D<
<D=
$
obtaining
, a ?sparse? version of the ridge regression estimator [12] (the
sparsity is due to the fact that we only store in the forwarded instances, i.e., those for
which we have a relevance labels). To estimate directly the margin , rather than ,
we further modify, along the lines of the techniques analyzed in [3, 6, 15], the sparse ridge
with the quantity %
, where
regression estimator. More precisely, we estimate
the % is defined by
<
K < %$ < % $ $ < %$ = %$
(3)
Using the Sherman-Morrison
can then write out the expectation of as
$
,formula,
we
F EI
,
which holds for all , , and all matrices
< %$ . Let % $ be the number of forwarded instances among % $ . In order to
generalize to the estimator (3) the analysis of
- , we need to find a large deviation
bound of the form
' , % $ ?
for all , , where
goes to zero ?sufficiently fast? as 4 . Though we have not been able to find such
%
4%
%
SIMPLE FIL
-8$BD %
A+
D
)
>
>
bounds, we report some experimental results showing that algorithms based on (3) and
inspired by the analysis of SIMPLE - FIL do exhibit a good empirical behavior on real-world
data. Moreover, in Section 4.2 we prove a bound (not based on the analysis of SIMPLE - FIL)
on the expected regret of a randomized variant of the algorithm used in the experiments.
For this variant we are able to prove a regret bound that scales essentially with the square
root of (to be contrasted with the logarithmic regret of SIMPLE - FIL).
4.1 Experimental results
We ran our experiments using the filtering rule that forwards
if SGN ;%
,
8'
where % is the estimator (3) and '
"! # %$ Note that this rule, which we call
RIDGE - FIL , is a natural generalization of SIMPLE - FIL to the unrestricted learning model; in
particular, SIMPLE - FIL uses a relevance threshold ' of the very same form as RIDGE - FIL,
although SIMPLE - FIL?s ?margin? ,
is defined differently. We tested our algorithm on a
, . "K< , .
.
1. Get and let
2. If
then forward , get label and update as follows:
$ ;
< <
$ < < ;
K , where if and is such
, otherwise;
that
.
3. Else forward with probability . If was forwarded then get label and do
the same updates as in 2; otherwise, do not make any update.
Algorithm: P - RIDGE - FIL.
Parameters: Real : ;
Initialization: % E4
Loop for
"
A
B5
9%
,+
,+
9
%
%
%
E
D D%
DD
@
,+
<%
D D%
3D D =
9
Figure 2: Pseudo-code for the filtering algorithm P - RIDGE - FIL. The performance of this
algorithm is analyzed in Theorem 3.
document filtering problem based on the first 70000 newswire stories from the Reuters
Corpus Volume 1. We selected the 34 Reuters topics whose frequency in the set of 70000
documents was between 1% and 5% (a plausible range for filtering applications). For
each topic, we defined a filtering task whose relevance judgements were assigned based
on whether the document was labelled with that topic or not. Documents were mapped
to real vectors using the bag-of-words representation. In particular, after tokenization we
lemmatized the tokens using a general-purpose finite-state morphological English analyzer
and then removed stopwords (we also replaced all digits with a single special character).
Document vectors were built removing all words which did not occur at least three times in
the corpus and using the TF-IDF encoding in the form
"! TF '(!7 DF , where TF is
the word frequency in the document, DF is the number of documents containing the word,
and is the total number of documents (if TF the TF-IDF coefficient was also set to
). Finally, all document vectors were normalized to length 1. To measure how the choice
of the threshold ' affects the filtering performance, we ran RIDGE - FIL with ' set to zero
on the same dataset as a standard on-line binary classifier (i.e., receiving the correct label
after every classification). We call this algorithm RIDGE - FULL. Figure 1 illustrates the
experimental results. The average -measure of RIDGE - FULL and RIDGE - FIL are, respectively,
and ; hence the threshold compensates pretty well the partial feedback in
the filtering setup. On the other hand, the standard Perceptron achieves here a -measure
in the classification task, hence inferior to that of RIDGE - FULL. Finally, we also
of
tested the apple-tasting filtering rule (see [9, STAP transformation]) based on the binary
classifier RIDGE - FULL. This transformation, which does not take into consideration the
margin, exhibited a poor performance and we did not include it in the plot.
4.2 Probabilistic ridge filtering
In this section we introduce a probabilistic filtering algorithm, derived from the (on-line)
ridge regression algorithm, for the class of linear probabilistic relevance functions. The
algorithm, called P - RIDGE - FIL, is sketched in Figure 2. The algorithm takes and a
probability value as input parameters and maintains a linear hypothesis % . If %
, then
is forwarded and %
gets updated according to the following two-steps ridge
regression-like rule. First, the intermediate vector %
is computed via the standard on-line
ridge regression algorithm using the inverse of matrix
. Then, the new vector %
is obtained by projecting %
onto the unit ball, where the projection is taken w.r.t. the
"
distance function
implies
; B%
4 (%
; (% . Note that D D % /D D =
%
%
. On the other hand, if %
0 then
is forwarded (and consequently
%
is updated) with some probability . The analysis of P - RIDGE - FIL is inspired by the
analysis in [1] for a related but different problem, and is based on relating the expected
regret in a given trial to a measure of the progress of % towards . The following lemma
will be useful.
Lemma 2 Using the notation of Figure 2, let be the trial when the -th update
< oc
<
= 8(!
curs. Then the following inequality holds:
,+
:
&+
:
; %
and
; B%
, where D /D denotes the determinant of matrix
; %
4 (%"
; 8% .
;
!
Theorem 3 Let !
. For all , if algorithm - ! ! , then its expected cumulative regret
of Figure 2 is run with
,+
, + is at most
!
!
" " " (! - #
Proof sketch. If is the trial when the -th forward takes place,
we define the random
& and ' 8 (! < < . If no update occurs
variables $ %
' . Let ) be the regret of - - in trial and ) be
in trial we set $ (
in trial . If
, then ) ) and
the regret of the update rule
$ can be lower bounded via Lemma 2. If
, then $ gets lower bounded via
Lemma 2 only with probability , while for the regret we can only use the trivial bound
)
. With probability 4 , instead, ) ) and $ . Let * be a
constant to be specified. We can write
) +
* $ , ) + * $
)
- * $
+
(4)
Now, it is easy to verify that in the conditional space where is given we have
and
. Thus, using Lemma
2 and Eq. (4) we can write
'
) +
* $ . ) + *
)
'
*6
Proof sketch. From Lemma 4.2 and Theorem 4.6 in [3] and the fact that D ,+ D = D D % D D =
<
= 8(!
4 B%
.
it follows that ,+
+
< ; B%
4
%
%" is a Bregman diverNow, the function
;
; B%
gence (e.g., [4, 10]), and it can be easily shown that %
in Figure 2 is the projection of %
onto the convex set
D D D D =
w.r.t.
; i.e., %
!
= =
%
.
By
a
projection
property
of
Bregman
divergences
(see,
e.g., the appendix in [10]) it follows that
4 B%
/
24 %
for all such
that D D D D = . Putting together gives the desired inequality.
D+
D
1
-.&
,+
012
-.&
+
%
D
D
$
;
B%
)
=
0
E
,+
,+
6
P RIDGE FIL
%
?
%
P RIDGE FIL
092
;
,+
!
,+
09>
6
6
:C
?!
,+
1>
!
,+
09>
! &+
,+
*
D ,+
?!
+
,+
D ,+
6 ?,+
1C
5
,+
6 ,+
+
+
,+
,+
A+
01C
(
+
C
D4,+
(
,+
;?,+
D4,+
;?,+
:C
0:C
01C
1
This parametrization requires the knowledge of / . It turns out one can remove this assumption
at the cost of a slightly more involved proof.
+ *
+*
* '
* '
+
(5)
where in the inequality we have dropped from factor and combined the resulting
and ) into ) . In turn, this term has been bounded
terms )
" by virtue of (2) with . At this
as )
point we work in the conditional space where
is given and distinguish the two cases
and . In the first case we have
- * * ' + * * '
whereas in the second case we have
* , * '
- ' # * * '
+ '
. We set *
where in both cases we used
sum over .
and
-
-
<
Notice that , + $
and that , + ' 8 "! < (! -
(e.g., [3], proof of Theorem 4.6 therein). After a few overapproximations (and taking the
and
) we obtain
worst between the two cases
#
#
, +
)
" "! - " #
"
This can be further upper bounded by
,+
A+
5D ,+
& ,+
,+
+
9>
?,+
!
1C
2D ,+
& ,+
,+
.01C
+
,+
,+
09>
01C
?
6 ,+
6 =
01C
6 ,+
,+
+
9>
=
6
,+
+
'
,+
,+
9
01
,+
1 ,+
2 9 ,+
+
A+
D ,+
D4,+
*
*
=
D D
?
C!
C
=
thereby concluding the proof.
1
=
=
=
&
?
>
?
D4,+
C
D ,+
=
=
$
)
01
,+
D4,+
D ,+
DD
,+
3+
,+
2
$
)
;
References
[1] Abe, N., and Long, P.M. (1999). Associative reinforcement learning using linear probabilistic
concepts. In Proc. ICML?99, Morgan Kaufmann.
[2] Auer, P. (2000). Using Upper Confidence Bounds for Online Learning. In Proc. FOCS?00,
IEEE, pages 270?279.
[3] Azoury, K., and Warmuth, M.K. (2001). Relative loss bounds for on-line density estimation
with the exponential family of distributions, Machine Learning, 43:211?246.
[4] Censor, Y., and Lent, A. (1981). An iterative row-action method for interval convex programming. Journal of Optimization Theory and Applications, 34(3), 321?353.
[5] Cesa-Bianchi, N. (1999). Analysis of two gradient-based algorithms for on-line regression.
Journal of Computer and System Sciences, 59(3):392?411.
[6] Cesa-Bianchi, N., Conconi, A., and Gentile, C. (2002). A second-order Perceptron algorithm.
In Proc. COLT?02, pages 121?137. LNAI 2375, Springer.
[7] Cesa-Bianchi, N., Long, P.M., and Warmuth, M.K. (1996). Worst-case quadratic loss bounds
for prediction using linear functions and gradient descent. IEEE Trans. NN, 7(3):604?619.
[8] Gavald`a, R., and Watanabe, O. (2001). Sequential sampling algorithms: Unified analysis and
lower bounds. In Proc. SAGA?01, pages 173?187. LNCS 2264, Springer.
[9] Helmbold, D.P., Littlestone, N., and Long, P.M. (2000). Apple tasting. Information and Computation, 161(2):85?139.
[10] Herbster, M. and Warmuth, M.K. (1998). Tracking the best regressor, in Proc. COLT?98, ACM,
pages 24?31.
[11] Hoeffding, W. (1963). Probability inequalities for sums of bounded random variables. Journal
of the American Statistical Association, 58:13?30.
[12] Hoerl, A., and Kennard, R. (1970). Ridge regression: biased estimation for nonorthogonal problems. Technometrics, 12:55?67.
[13] Vapnik, V. (1998). Statistical learning theory. New York: J. Wiley & Sons.
[14] Voorhees, E., Harman, D. (2001). The tenth Text REtrieval Conference. TR 500-250, NIST.
[15] Vovk, V. (2001). Competitive on-line statistics. International Statistical Review, 69:213?248.
| 2154 |@word trial:6 determinant:1 version:3 judgement:6 thereby:1 minus:1 tr:1 document:13 ours:1 bc:2 current:1 must:1 bd:3 happen:1 remove:2 drop:1 plot:2 update:6 selected:1 warmuth:3 parametrization:1 dell:1 stopwords:1 along:1 focs:1 prove:5 introduce:2 indeed:1 expected:7 behavior:2 inspired:3 actual:1 project:1 moreover:1 notation:1 bounded:5 kind:2 developed:1 unified:1 transformation:4 pseudo:3 dti:4 every:1 unwanted:1 universit:1 classifier:5 wrong:2 control:1 unit:1 superiority:1 positive:3 dropped:2 crema:2 modify:1 mistake:5 crii:1 encoding:1 analyzing:1 nonrelevant:3 might:2 tasting:5 initialization:1 therein:1 dynamically:1 forwarding:2 range:1 regret:17 digit:1 lncs:1 empirical:4 adapting:1 projection:3 confidence:4 fraud:1 integrating:1 word:4 suggest:1 get:8 cannot:1 close:1 onto:2 applying:2 go:2 convex:2 simplicity:1 helmbold:1 rule:25 estimator:8 insubria:1 fulfil:1 updated:3 target:1 user:3 programming:1 us:1 hypothesis:2 element:6 observed:1 worst:2 news:1 momentarily:1 morphological:1 removed:1 observes:2 ran:2 balanced:1 easily:2 differently:1 fast:2 kermit:1 whose:3 larger:1 valued:2 plausible:1 say:3 otherwise:3 forwarded:13 compensates:1 statistic:1 itself:1 associative:1 obviously:1 online:1 sequence:6 relevant:10 loop:1 realization:1 achieve:1 milan:2 derive:1 progress:1 eq:1 implies:3 convention:1 correct:3 filter:4 sgn:2 implementing:1 everything:1 require:1 piecing:1 suffices:1 fix:1 generalization:1 preliminary:3 hold:4 fil:32 sufficiently:1 nonorthogonal:1 predict:1 achieves:1 purpose:1 estimation:2 proc:5 bag:1 label:13 hoerl:1 individually:1 tf:5 always:1 rather:2 claudio:1 probabilistically:1 derived:1 notational:1 consistently:1 bernoulli:1 rank:1 censor:1 nn:1 lnai:1 hidden:2 sketched:1 overall:1 classification:4 among:4 colt:2 special:1 tokenization:1 aware:1 never:2 sampling:1 bramante:2 chernoff:2 adversarially:1 look:1 icml:1 report:2 few:1 divergence:1 replaced:1 intended:1 maintain:1 technometrics:1 analyzed:2 bregman:2 partial:2 littlestone:1 desired:2 theoretical:1 instance:7 column:2 cost:1 deviation:3 uninteresting:1 predicate:1 too:1 commission:1 combined:1 density:1 herbster:1 randomized:4 international:1 probabilistic:9 receiving:1 regressor:1 together:2 quickly:1 again:1 cesa:5 opposed:1 containing:1 hoeffding:3 american:1 account:1 coefficient:1 matter:1 stream:2 piece:1 view:1 root:1 start:1 competitive:1 maintains:1 square:4 variance:1 kaufmann:1 generalize:3 apple:5 classified:2 ed:1 email:1 definition:1 frequency:3 involved:1 associated:2 proof:6 cur:1 dataset:1 recall:3 knowledge:1 formalize:1 auer:1 higher:1 though:1 generality:1 sketch:3 hand:2 lent:1 ei:1 building:1 verify:2 true:2 unbiased:1 normalized:1 concept:1 hence:5 equality:1 assigned:1 harman:1 inferior:1 oc:1 d4:5 criterion:3 ridge:34 reasoning:1 consideration:1 ji:1 volume:2 association:1 relating:1 similarly:1 newswire:1 analyzer:1 sherman:1 italy:3 store:1 certain:1 inequality:5 binary:3 morgan:1 gentile:3 unrestricted:4 employed:1 converge:1 morrison:1 full:9 characterized:1 adapt:1 believed:1 long:3 retrieval:2 prediction:2 variant:5 regression:8 essentially:1 expectation:1 df:2 kernel:1 whereas:2 want:2 separately:1 interval:1 else:1 crucial:1 biased:1 unlike:1 exhibited:1 call:5 revealed:1 intermediate:1 enough:2 easy:1 affect:1 knowing:1 whether:3 york:1 cause:1 action:2 useful:1 category:1 generate:1 notice:1 judging:1 sign:1 discrete:1 write:4 putting:1 threshold:6 changing:1 varese:1 tenth:1 fraction:3 sum:3 run:2 inverse:4 logged:1 place:1 almost:2 family:1 decide:2 appendix:1 bound:20 distinguish:1 quadratic:1 adapted:1 dangerous:1 occur:1 precisely:1 idf:2 alex:1 argument:1 concluding:1 according:3 ball:1 poor:1 belonging:1 remain:1 slightly:2 son:1 character:1 making:1 projecting:1 restricted:3 taken:1 remains:1 turn:2 know:2 observe:3 original:1 denotes:1 include:1 quantity:1 occurs:1 exhibit:1 gradient:2 distance:1 mapped:1 gence:1 topic:4 trivial:1 code:1 length:1 index:1 setup:1 negative:3 design:2 unknown:7 perform:2 bianchi:5 imbalance:1 upper:2 observation:1 markov:1 finite:2 nist:1 descent:1 arbitrary:2 community:1 drift:1 abe:1 pair:1 specified:1 trans:1 able:3 pattern:1 sparsity:1 built:1 suitable:1 event:1 natural:1 irrespective:1 gf:1 text:1 review:1 nicol:1 relative:1 loss:3 dsi:1 filtering:36 proven:1 dd:2 story:1 cd:19 row:1 token:1 supported:1 last:1 english:1 guide:1 perceptron:2 taking:1 sparse:2 feedback:2 world:2 cumulative:2 forward:13 made:2 adaptive:1 reinforcement:1 simplified:1 spam:1 counted:1 decides:1 corpus:3 assumed:1 iterative:1 pretty:1 obtaining:1 european:1 protocol:1 did:2 azoury:1 reuters:3 scored:1 allowed:1 kennard:1 wiley:1 precision:3 watanabe:1 saga:1 exponential:1 theorem:5 formula:1 e4:1 removing:1 showing:1 voorhees:1 virtue:1 exists:1 false:5 adding:1 sequential:1 importance:1 vapnik:1 illustrates:1 margin:9 logarithmic:2 likely:1 conconi:3 tracking:1 springer:2 complemented:1 acm:1 conditional:3 viewed:1 identity:1 consequently:1 towards:1 labelled:2 replace:1 contrasted:1 unimi:3 vovk:1 lemma:6 called:2 total:1 experimental:4 relevance:18 tested:2 |
1,268 | 2,155 | Independent Components Analysis
through Product Density Estimation
'frevor Hastie and Rob Tibshirani
Department of Statistics
Stanford University
Stanford, CA, 94305
{ hastie, tibs } @stat.stanford. edu
Abstract
We present a simple direct approach for solving the ICA problem,
using density estimation and maximum likelihood. Given a candidate orthogonal frame, we model each of the coordinates using a
semi-parametric density estimate based on cubic splines. Since our
estimates have two continuous derivatives , we can easily run a second order search for the frame parameters. Our method performs
very favorably when compared to state-of-the-art techniques.
1
Introduction
Independent component analysis (ICA) is a popular enhancement over principal
component analysis (PCA) and factor analysis. In its simplest form , we observe a
random vector X E IRP which is assumed to arise from a linear mixing of a latent
random source vector S E IRP,
(1)
X=AS;
the components Sj, j = 1, ... ,p of S are assumed to be independently distributed.
The classical example of such a system is known as the "cocktail party" problem.
Several people are speaking, music is playing, etc., and microphones around the
room record a mix of the sounds. The ICA model is used to extract the original
sources from these different mixtures.
Without loss of generality, we assume E(S) = 0 and Cov(S) = I , and hence
Cov(X) = AA T . Suppose S* = R S represents a transformed version of S, where R
is p x p and orthogonal. Then with A * = ART we have X* = A * S* = ARTR S =
X. Hence the second order moments Cov(X) = AAT = A * A *T do not contain
enough information to distinguish these two situations.
Model (1) is similar to the factor analysis model (Mardia, Kent & Bibby 1979),
where S and hence X are assumed to have a Gaussian density, and inference is
typically based on the likelihood of the observed data. The factor analysis model
typically has fewer than p components, and includes an error component for each
variable. While similar modifications are possible here as well, we focus on the
full-component model in this paper. Two facts are clear:
? Since a multivariate Gaussian distribution is completely determined by its
first and second moments, this model would not be able to distinguish A
and A * . Indeed, in factor analysis one chooses from a family of factor
rotations to select a suitably interpretable version.
? Multivariate Gaussian distributions are completely specified by their
second-order moments. If we hope to recover the original A, at least p - 1
of the components of S will have to be non-Gaussian.
Because of the lack of information in the second moments, the first step in an ICA
model is typically to transform X to have a scalar covariance, or to pre-whiten the
data. From now on we assume Cov(X) = I , which implies that A is orthogonal.
Suppose the density of Sj is Ij, j = 1, ... ,p, where at most one of the Ii are
Gaussian. Then the joint density of S is
p
(2)
Is(s) =
II Ii(Sj),
j= l
and since A is orthogonal, the joint density of X is
p
(3)
Ix(x) =
II Ii(aJ x),
j=l
where aj is the jth column of A . Equation (3) follows from S = AT X due to the
orthogonality of A , and the fact that the determinant in this multivariate transformation is 1.
In this paper we fit the model (3) directly using semi-parametric maximum likelihood. We represent each of the densities Ii by an exponentially tilted Gaussian
density (Efron & Tibshirani 1996).
(4)
where ? is the standard univariate Gaussian density, and gj is a smooth function,
restricted so that Ii integrates to 1. We represent each of the functions gj by a cubic
smoothing spline, a rich class of smooth functions whose roughness is controlled by a
penalty functional. These choices lead to an attractive and effective semi-parametric
implementation of ICA:
? Given A, each of the components Ii in (3) can be estimated separately by
maximum likelihood. Simple algorithms and standard software are available.
? The components gj represent departures from Gaussianity, and the expected log-likelihood ratio between model (3) and the gaussian density is
given by Ex 2: j gj(aJ X), a flexible contrast function.
? Since the first and second derivatives of each of the estimated gj are immediately available, second order methods are available for estimating the
orthogonal matrix A . We use the fixed point algorithms described in
(Hyvarinen & Oja 1999).
? Our representation of the gj as smoothing splines casts the estimation problem as density estimation in a reproducing kernel Hilbert space, an infinite
family of smooth functions. This makes it directly comparable with the
"Kernel ICA" approach of Bach & Jordan (2001), with the advantage that
we have O(N) algorithms available for the computation of our contrast
function, and its first two derivatives.
In the remainder of this article, we describe the model in more detail, and evaluate
its performance on some simulated data.
2
Fitting the Product Density leA model
Given a sample Xl, ... ,XN we fit the model (3),(4) by maximum penalized likelihood. The data are first transformed to have zero mean vector, and identity
covariance matrix using the singular value decomposition. We then maximize the
criterion
(5)
subject to
T
(6)
J
a j ak
bjk
?(s)e 9j (slds
(7)
't/j, k
1 't/j
For fixed aj and hence Sij = aT Xi the solutions for 9j are known to be cubic splines
with knots at each of the unique values of Sij (Silverman 1986). The p terms
decouple for fixed aj, leaving us p separate penalized density estimation problems.
We fit the functions 9j and directions aj by optimizing (5) in an alternating fashion ,
as described in Algorithm 1. In step (a), we find the optimal 9j for fixed 9j; in
Algorithm 1 Product Density leA algorithm
1. Initialize A (random Gaussian matrix followed by orthogonalization).
2. Alternate until convergence of A, using the Amari metric (16).
(a) Given A , optimize (5) w.r.t. 9j (separately for each j), using the
penalized density estimation algorithm 2.
(b) Given 9j , j = 1, ... ,p, perform one step of the fixed point algorithm 3
towards finding the optimal A.
step (b), we take a single fixed-point step towards the optimal A. In this sense
Algorithm 1 can be seen to be maximizing the profile penalized log-likelihood w.r.t.
A.
2.1
Penalized density estimation
We focus on a single coordinate, with N observations Si,
Si = Xi for some k). We wish to maximize
1, ... ,N (where
af
(8)
J
subject to ?(s)e 9 (slds = 1. Silverman (1982) shows that one can incorporate the
integration constraint by using the modified criterion (without a Lagrange multiplier)
N
(9)
~ l:= {lOg?(Si) + 9(Si )} >=1
J
?(s)e 9 (slds - A
J
91/ 2 (S)ds.
Since (9) involves an integral, we need an approximation. We construct a fine grid
of L values s; in increments ~ covering the observed values Si, and let
*
(10)
Y? =
#Si E (sf - ~/2, Sf
N
+ ~/2)
Typically we pick L to be 1000, which is more than adequate. We can then approximate (9) by
L
(11)
L
{Y; [log(?(s;))
+ g(s;)]- ~?(se)e9(sll}
J
- A
gI/2(s)ds.
?=1
This last expression can be seen to be proportional to a penalized Poisson loglikelihood with response Y;! ~ and penalty parameter A/~, and mean J-t(s) =
?(s)e 9(s). This is a generalized additive model (Hastie & Tibshirani 1990), with
an offset term log(?(s)), and can be fit using a Newton algorithm in O(L) operations. As with other GAMs, the Newton algorithm is conveniently re-expressed as
an iteratively reweighted penalized least squares regression problem, which we give
in Algorithm 2.
Algorithm 2 Iteratively reweighted penalized least squares algorithm for fitting
the tilted Gaussian spline density model.
1. Initialize 9 == O.
2. Repeat until convergence:
(a) Let J-t(s;) = ?(s;)e 9(sll, ? = 1, ... ,L, and w?
(b) Define the working response
(12)
z? = g(s*)
?
= J-t(s;).
+ Ye -
J-t(sf)
J-t( sf)
(c) Update g by solving the weighted penalized least squares problem
(13)
This amounts to fitting a weighted smoothing spline to the pairs (sf, ze)
with weights w? and tuning parameter 2A/~.
Although other semi-parametric regression procedures could be used in (13), the
cubic smoothing spline has several advantages:
? It has knots at all L of the pseudo observation sites sf' The values sf
can be fixed for all terms in the model (5), and so a certain amount of
pre-computation can be performed. Despite the large number of knots
and hence basis functions , the local support of the B-spline basis functions
allows the solution to (13) to be obtained in O(L) computations.
? The first and second derivatives of 9 are immediately available, and are
used in the second-order search for the direction aj in Algorithm 1.
? As an alternative to choosing a value for A, we can control the amount of
smoothing through the effective number of parameters, given by the trace
of the linear operator matrix implicit in (13) (Hastie & Tibshirani 1990).
? It can also be shown that because of the form of (9), the resulting density
inherits the mean and variance of the data (0 and 1); details will be given
in a longer version of this paper.
2.2
A fixed point method for finding the orthogonal frame
For fixed functions g1> the penalty term in (5) does not playa role in the search
for A. Since all of the columns aj of any A under consideration are mutually
orthogonal and unit norm, the Gaussian component
p
L log ?(aJ
Xi)
j=l
does not depend on A. Hence what remains to be optimized can be seen as the
log-likelihood ratio between the fitted model and the Gaussian model, which is
simply
C(A)
(14)
Since the choice of each gj improves the log-likelihood relative to the Gaussian, it is
easy to show that C(A) is positive and zero only if, for the particular value of A, the
log-likelihood cannot distinguish the tilted model from a Gaussian model. C(A) has
the form of a sum of contrast functions for detecting departures from Gaussianity.
Hyvarinen, Karhunen & Oja (2001) refer to the expected log-likelihood ratio as the
negentropy, and use simple contrast functions to approximate it in their FastICA
algorithm. Our regularized approach can be seen as a way to construct a flexible
contrast function adaptively using a large set of basis functions .
Algorithm 3 Fixed point update forA.
1. For j = 1, ... ,p:
(15)
where E represents expectation w.r.t. the sample
column of A.
Xi,
and
aj
is the jth
2. Orthogonalize A: Compute its SVD , A = UDV T , and replace A
f-
UV T .
Since we have first and second derivatives avaiable for each gj , we can mimic exactly
the fast fixed point algorithm developed in (Hyvarinen et al. 2001, page 189) ; see
algorithm 3. Figure 1 shows the optimization criterion C (14) above, as well as the
two criteria used to approximate negentropy in FastICA by Hyvarinen et al. (2001)
[page 184]. While the latter two agree with C quite well for the uniform example
(left panel), they both fail on the mixture-of-Gaussians example, while C is also
successful there.
Uniforms
x
Gaussian Mixtures
0
0
'",,;
'",,;
'",,;
x
"
'",,;
"
"0
C
"0
C
'",,;
'",,;
'",,;
'",,;
0
0
,,;
,,;
0.0
0.5
1.0
1.5
e
2.0
2.5
3.0
0.0
0.5
1.0
1.5
e
2.0
2.5
3.0
Figure 1: The optimization criteria and solutions found for two different examples in lR2
using FastICA and our ProDenICA . G 1 and G2 refer to the two functions used to define
negentropy in FastICA. In the left example the independent components are uniformly
distributed, in the right a mixture of Gaussians. In the left plot, all the procedures found
the correct frame; in the right plot, only the spline based approach was successful. The
vertical lines indicate the solutions found, and the two tick marks at the top of each plot
indicate the true angles.
3
Comparisons with fast ICA
In this section we evaluate the performance of the product density approach (ProDenICA) , by mimicking some of the simulations performed by Bach & Jordan (2001)
to demonstrate their Kernel ICA approach. Here we compare ProDenICA only with
FastICA; a future expanded version of this paper will include comparisons with other
leA procedures as well.
The left panel in Figure 2 shows the 18 distributions used as a basis of comparison.
These exactly or very closely approximate those used by Bach & Jordan (2001) . For
each distribution, we generated a pair of independent components (N=1024) , and
a random mixing matrix in ill? with condition number between 1 and 2. We used
our Splus implementation of the FastICA algorithm, using the negentropy criterion
based on the nonlinearity G 1 (s) = log cosh(s) , and the symmetric orthogonalization
scheme as in Algorithm 3 (Hyvarinen et al . 2001, Section 8.4.3). Our ProDenICA
method is also implemented in Splus. For both methods we used five random starts
(without iterations). Each of the algorithms delivers an orthogonal mixing matrix A
(the data were pre-whitenea) , which is available for comparison with the generating
orthogonalized mixing matrix A o. We used the Amari metric(Bach & Jordan 2001)
as a measure of the closeness of the two frames:
(16)
d(Ao,A) =
~ f.- (L~=1 Irijl
2p ~
i=1
max?lr?
?1
J"J
-1) + ~ f.- (Lf=1I rijl -1) ,
2p ~
max?lr??1
j=1"
"J
where rij = (AoA - 1 )ij . The right panel in Figure 2 shows boxplots of the pairwise
differences d(Ao, A F ) -d(Ao , Ap ) (x100), where the subscripts denote ProDenICA
or FastICA. ProDenICA is competitive with FastICA in all situations, and dominates in most of the mixture simulations. The average Amari error (x 100) for
FastICA was 13.4 (2.7), compared with 3.0 (0.4) for ProDenICA (Bach & Jordan
(2001) report averages of 6.2 for FastICA, and 3.8 and 2.9 for their two KernelICA
methods).
We also ran 300 simulations in
1R.4,
using N = 1000, and selecting four of the
,
b
,
0 ..,---_ _ _ _ _ _ _ _ _ _ _ _ _----,
JL ~
d
~
9
~
~- JJL
f
h
flL ~ ~
~~~
j
k
m
"
I
,
~~~
p
ro
ci
q
~~~
N
ci
o
ci
abcdefghijklmnopqr
distribution
Figure 2: The left panel shows eighteen distributions used for comparisons. These include
the "t", uniform, exponential, mixtures of exponentials, symmetric and asymmetric gaussian mixtures. The right panel shows boxplots of the improvement of ProDenICA over
FastICA in each case, using the Amari metric, based on 30 simulations in lR? for each
distribution.
18 distributions at random. The average Amari error (x 100) for FastICA was
26.1 (1.5), compared with 9.3 (0.6) for ProDenICA (Bach & Jordan (2001) report
averages of 19 for FastICA , and 13 and 9 for their two K ernelICA methods).
4
Discussion
The lCA model stipulates that after a suitable orthogonal transformation, the data
are independently distributed. We implement this specification directly using semiparametric product-density estimation. Our model delivers estimates of both the
mixing matrix A, and estimates of the densities of the independent components.
Many approaches to lCA, including FastICA, are based on minimizing approximations to entropy. The argument, given in detail in Hyvarinen et al. (2001) and
reproduced in Hastie, Tibshirani & Friedman (2001), starts with minimizing the
mutual information - the KL divergence between the full density and its independence version. FastICA uses very simple approximations based on single (or a small
number of) non-linear contrast functions , which work well for a variety of situations,
but not at all well for the more complex gaussian mixtures. The log-likelihood for
the spline-based product-density model can be seen as a direct estimate of the mutual information; it uses the empirical distribution of the observed data to represent
their joint density, and the product-density model to represent the independence
density. This approach works well in both the simple and complex situations automatically, at a very modest increase in computational effort. As a side benefit,
the form of our tilted Gaussian density estimate allows our log-likelihood criterion
to be interpreted as an estimate of negentropy, a measure of departure from the
Gaussian.
Bach & Jordan (2001) combine a nonparametric density approach (via reproducing
kernel Hilbert function spaces) with a complex measure of independence based on
the maximal correlation. Their procure requires O(N3) computations, compared to
our O(N). They motivate their independence measures as approximations to the
mutual independence. Since the smoothing splines are exactly function estimates
in a RKHS, our method shares this flexibility with their Kernel approach (and is
in fact a "Kernel" method). Our objective function, however, is a much simpler
estimate of the mutual information. In the simulations we have performed so far ,
it seems we achieve comparable accuracy.
References
Bach, F . & Jordan, M. (2001), Kernel independent component analysis, Technical
Report UCBjCSD-01-1166, Computer Science Division, University of California, Berkeley.
Efron, B. & Tibshirani, R. (1996), 'Using specially designed exponential families
for density estimation' , Annals of Statistics 24(6), 2431-246l.
Hastie, T. & Tibshirani, R. (1990), Generalized Additive Models, Chapman and
Hall.
Hastie, T., Tibshirani, R. & Friedman, J. (2001), The Elements of Statistical Learning; Data mining, Inference and Prediction, Springer Verlag, New York.
Hyvarinen, A., Karhunen, J. & Oja, E. (2001), Independent Component Analysis,
Wiley, New York.
Hyvarinen, A. & Oja, E. (1999), 'Independent component analysis: Algorithms and
applications' , Neural Networks .
Mardia, K., Kent, J. & Bibby, J. (1979), Multivariate Analysis, Academic Press.
Silverman, B. (1982), 'On the estimation of a probability density function by the
maximum penalized likelihood method', Annals of Statistics 10(3),795-810.
Silverman, B. (1986), Density Estimation for Statistics and Data Analysis, Chapman and Hall.
| 2155 |@word determinant:1 version:5 norm:1 seems:1 suitably:1 simulation:5 covariance:2 kent:2 decomposition:1 pick:1 moment:4 selecting:1 rkhs:1 si:6 negentropy:5 tilted:4 additive:2 plot:3 interpretable:1 update:2 designed:1 fewer:1 record:1 lr:3 detecting:1 simpler:1 five:1 direct:2 fitting:3 combine:1 pairwise:1 indeed:1 expected:2 ica:8 udv:1 automatically:1 estimating:1 panel:5 what:1 interpreted:1 developed:1 finding:2 transformation:2 pseudo:1 berkeley:1 exactly:3 ro:1 control:1 unit:1 positive:1 aat:1 local:1 despite:1 ak:1 subscript:1 ap:1 bjk:1 unique:1 implement:1 lf:1 silverman:4 procedure:3 empirical:1 pre:3 cannot:1 operator:1 optimize:1 maximizing:1 independently:2 immediately:2 coordinate:2 increment:1 annals:2 suppose:2 us:2 element:1 ze:1 asymmetric:1 observed:3 role:1 tib:1 rij:1 ran:1 motivate:1 depend:1 solving:2 division:1 completely:2 basis:4 easily:1 joint:3 x100:1 fast:2 effective:2 describe:1 choosing:1 whose:1 quite:1 stanford:3 loglikelihood:1 amari:5 statistic:4 cov:4 gi:1 g1:1 transform:1 reproduced:1 advantage:2 product:7 maximal:1 remainder:1 sll:2 mixing:5 flexibility:1 achieve:1 convergence:2 enhancement:1 generating:1 stat:1 ij:2 implemented:1 involves:1 implies:1 indicate:2 direction:2 closely:1 correct:1 ao:3 roughness:1 around:1 hall:2 estimation:11 integrates:1 weighted:2 hope:1 gaussian:19 modified:1 focus:2 inherits:1 improvement:1 likelihood:14 contrast:6 sense:1 inference:2 irp:2 typically:4 transformed:2 mimicking:1 flexible:2 ill:1 art:2 smoothing:6 initialize:2 integration:1 jjl:1 mutual:4 construct:2 chapman:2 represents:2 mimic:1 future:1 report:3 spline:11 oja:4 divergence:1 friedman:2 mining:1 mixture:8 integral:1 orthogonal:9 modest:1 re:1 orthogonalized:1 fitted:1 column:3 bibby:2 uniform:3 fastica:15 successful:2 chooses:1 adaptively:1 density:32 e9:1 derivative:5 includes:1 gaussianity:2 performed:3 start:2 recover:1 competitive:1 lr2:1 square:3 accuracy:1 variance:1 knot:3 popular:1 efron:2 improves:1 hilbert:2 response:2 generality:1 implicit:1 until:2 d:2 working:1 correlation:1 lack:1 aj:10 ye:1 contain:1 true:1 multiplier:1 hence:6 alternating:1 symmetric:2 iteratively:2 attractive:1 reweighted:2 covering:1 whiten:1 criterion:7 generalized:2 demonstrate:1 performs:1 delivers:2 orthogonalization:2 consideration:1 rotation:1 functional:1 exponentially:1 jl:1 refer:2 tuning:1 uv:1 grid:1 nonlinearity:1 specification:1 longer:1 gj:8 etc:1 playa:1 multivariate:4 optimizing:1 certain:1 verlag:1 seen:5 maximize:2 semi:4 ii:8 full:2 mix:1 sound:1 smooth:3 technical:1 academic:1 af:1 bach:8 controlled:1 prediction:1 regression:2 expectation:1 metric:3 poisson:1 iteration:1 represent:5 kernel:7 lea:3 semiparametric:1 separately:2 fine:1 singular:1 source:2 leaving:1 specially:1 subject:2 jordan:8 enough:1 easy:1 variety:1 independence:5 fit:4 hastie:7 expression:1 pca:1 effort:1 penalty:3 speaking:1 york:2 adequate:1 cocktail:1 clear:1 se:1 amount:3 nonparametric:1 cosh:1 simplest:1 estimated:2 tibshirani:8 stipulates:1 splus:2 four:1 boxplots:2 sum:1 run:1 angle:1 family:3 comparable:2 followed:1 distinguish:3 fll:1 orthogonality:1 constraint:1 n3:1 software:1 argument:1 expanded:1 department:1 lca:2 alternate:1 rob:1 modification:1 restricted:1 sij:2 equation:1 mutually:1 remains:1 agree:1 fail:1 available:6 operation:1 gaussians:2 observe:1 gam:1 alternative:1 original:2 top:1 include:2 newton:2 music:1 forum:1 classical:1 objective:1 parametric:4 separate:1 simulated:1 ratio:3 minimizing:2 favorably:1 trace:1 implementation:2 perform:1 vertical:1 observation:2 eighteen:1 situation:4 frame:5 reproducing:2 cast:1 pair:2 specified:1 kl:1 optimized:1 california:1 able:1 departure:3 max:2 including:1 suitable:1 regularized:1 scheme:1 extract:1 relative:1 loss:1 proportional:1 kernelica:1 article:1 playing:1 share:1 penalized:10 repeat:1 last:1 jth:2 tick:1 side:1 distributed:3 benefit:1 xn:1 rich:1 party:1 far:1 hyvarinen:8 sj:3 approximate:4 assumed:3 xi:4 continuous:1 search:3 latent:1 ca:1 complex:3 arise:1 profile:1 site:1 cubic:4 fashion:1 wiley:1 wish:1 sf:7 xl:1 candidate:1 exponential:3 mardia:2 ix:1 offset:1 closeness:1 dominates:1 ci:3 karhunen:2 slds:3 entropy:1 simply:1 univariate:1 conveniently:1 lagrange:1 expressed:1 aoa:1 g2:1 scalar:1 springer:1 aa:1 identity:1 towards:2 room:1 replace:1 determined:1 infinite:1 uniformly:1 decouple:1 principal:1 microphone:1 svd:1 orthogonalize:1 select:1 people:1 support:1 latter:1 mark:1 incorporate:1 evaluate:2 ex:1 |
1,269 | 2,156 | Adaptive Scaling for Feature Selection in SVMs
Yves Grandvalet
Heudiasyc, UMR CNRS 6599,
Universit?e de Technologie de Compi`egne,
Compi`egne, France
[email protected]
St?ephane Canu
PSI
INSA de Rouen,
St Etienne du Rouvray, France
[email protected]
Abstract
This paper introduces an algorithm for the automatic relevance determination of input variables in kernelized Support Vector Machines. Relevance is measured by scale factors defining the input space metric, and
feature selection is performed by assigning zero weights to irrelevant
variables. The metric is automatically tuned by the minimization of the
standard SVM empirical risk, where scale factors are added to the usual
set of parameters defining the classifier. Feature selection is achieved
by constraints encouraging the sparsity of scale factors. The resulting
algorithm compares favorably to state-of-the-art feature selection procedures and demonstrates its effectiveness on a demanding facial expression recognition problem.
1 Introduction
In pattern recognition, the problem of selecting relevant variables is difficult. Optimal
subset selection is attractive as it yields simple and interpretable models, but it is a combinatorial and acknowledged unstable procedure [2]. In some problems, it may be better
to resort to stable procedures penalizing irrelevant variables. This paper introduces such a
procedure applied to Support Vector Machines (SVM).
The relevance of input features may be measured by continuous weights or scale factors,
which define a diagonal metric in input space. Feature selection consists then in determining a sparse diagonal metric, and sparsity can be encouraged by constraining an appropriate
norm on scale factors. Our approach can be summarized by the setting of a global optimization problem pertaining to 1) the parameters of the SVM classifier, and 2) the parameters
of the feature space mapping defining the metric in input space. As in standard SVMs,
only two tunable hyper-parameters are to be set: the penalization of training errors, and
the magnitude of kernel bandwiths. In this formalism we derive an efficient algorithm to
monitor slack variables when optimizing the metric. The resulting algorithm is fast and
stable.
After presenting previous approaches to hard and soft feature selection procedures in the
context of SVMs, we present our algorithm. This exposure is followed by an experimental
section illustrating its performances and conclusive remarks.
2 Feature Selection via adaptive scaling
Scaling is a usual preprocessing step, which has important outcomes in many classification
methods including SVM classifiers [9, 3]. It is defined by a linear transformation within
the input space:
, where
diag
is a diagonal matrix
of scale
factors.
Adaptive scaling consists in letting to be adapted during the estimation process with the
explicit aim of achieving a better recognition rate. For kernel classifiers, is a set of hyperparameters of the learning process. According to the structural risk minimization principle
[8], can be tuned in two ways:
1. estimate the parameters of classifier by empirical risk minimization for several values of
to produce a structure of classifiers
multi-indexed by
. Select one element of the structure by finding the set
minimizing some estimate of generalization error.
2. estimate the parameters of classifier and the hyper-parameters
by empirical risk minimization, while a second level hyper-parameter, say , constrains
in order to avoid overfitting. This procedure produces a structure of classifiers indexed by , whose value is computed by minimizing some estimate of
generalization error.
!
"
!
"
%
#
$
!%
The usual paradigm consists in computing the estimate of generalization error for regularly
spaced hyper-parameter values and picking the best solution among all trials. Hence, the
first approach requires intensive computation, since the trials should be completed over a
-dimensional grid over
values.
&
'
Several authors suggested to address this problem by optimizing an estimate of generalization error with respect to the hyper-parameters. For SVM classifiers, Cristianini et al. [4]
first proposed to apply an iterative optimization scheme to estimate a single kernel width
hyper-parameter. Weston et al. [9] and Chapelle et al. [3] generalized this approach to
multiple hyper-parameters in order to perform adaptive scaling and variable selection.
The experimental results in [9, 3] show the benefits of this optimization. However, relying on the optimization of generalization error estimates over many hyper-parameters is
hazardous. Once optimized, the unbiased estimates become down-biased, and the bounds
provided by VC-theory usually hold for kernels defined a priori (see the proviso on the
radius/margin bound in [8]). Optimizing these criteria may thus result in overfitting.
%
#
!
In the second solution considered here, the estimate of generalization error is minimized
with respect to , a single (second level) hyper-parameter, which constrains
.
The role of this constraint is twofold: control the complexity of the classifier, and encourage variable selection in input space. This approach is related to some successful
soft-selection procedures, such as lasso and bridge [5] in the frequentist framework and
Automatic Relevance Determination (ARD) [7] in the Bayesian framework. Note that this
type of optimization procedure has been proposed for linear SVM in both frequentist [1]
and Bayesian frameworks [6]. Our method generalizes this approach to nonlinear SVM.
3 Algorithm
3.1 Support Vector Machines
(*),+.-$/ </ ? <A0@ , where<CB functionB
214365 "798 ;:=<2> 798
The decision function provided by SVM is
is defined as:
(1)
B
1
where the parameters
8
are obtained by solving the following optimization problem:
<
<
4
1
3
1
:
7
<
) -
<
< BB
(2)
>
1
5
3
<
subject to
798 BB
with 5 defined as 5 . In this problem setting,
and the parameters of the
feature space mapping (typically a kernel bandwidth) are tunable hyper-parameters which
need to be determined by the user.
B
1
In [9, 3], adaptive scaling is performed by iteratively finding the parameters
8 of the
SVM classifier for a fixed value of
!
and minimizing aB bound on the estimate of generalization error with respect to hyper-parameters
. The algorithm
3.2 A global optimization problem
minimizes 1) the SVM empirical criterion with respect to parameters and 2) an estimate of
generalization error with respect to hyper-parameters.
!
In the present approach, we avoid the enlargement of the set of hyper-parameters by letting
to be standard parameters of the classifier. Complexity is controlled by
and
by constraining the magnitude of . The latter defines the single hyper-parameter of the
learning process related to scaling variables. The learning criterion is defined as follows:
<
1
1
<
3
:
)- )-
7
"
!
> < < 1 3 5 < "
subject to
7 8
$
&
: " &%
&%
<
B #B
BB#
'
(3)
BB &
B
%
In (3), the constraint on should favor sparse solutions. To allow
to go to zero, ( should
be
To encourage sparsity, zeroing a small #
should allow a high increase of &) ,
*, + positive.
' , hence ( should be small. In the limit of (-. , the constraint counts the number
Like in standard SVM classification, the minimization of an estimate of generalization error
is postponed to a later step, which consists in picking the best solution among all trials on
.
the two dimensional grid of hyper-parameters
of non-zero scale parameters, resulting in a hard selection procedure. This choice might
seem appropriate for our purpose, but it amounts to attempt to solve a highly non-convex
optimization problem, where the number of local minima grows exponentially with the
input dimension . To avoid this problem, we suggest to use (
, which is the smallest
value for which the problem is convex with the linear mapping
. Indeed, for
linear kernels, the constraint on amounts to minimize the standard SVM criterion where
the penalization
on the /10 norm is replaced by the penalization of the /3254 norm. Hence,
provides
47682
setting (
the solution of the / SVM classifier described in [1]. For non-linear
kernels however, the two solutions differ notably since the present algorithm modifies the
metric in input space, while the / SVM classifier modifies
the metric in feature space.
and Gaussian
Finally, note that unicity can be guaranteed for (
kernels with large
bandwidths ( -9 ).
&
5
%
3.3 An alternated optimization scheme
B
1 8
Problem (3) is complex; we propose to solve iteratively a series of simplier problems.
The function is first optimized with respect to parameters
for a fixed mapping
(standard SVM problem). Then, the parameters of the feature space mapping are
optimized
while some characteristics
of
are kept fixed: At step , starting
from a given
value, the optimal
are computed. Then
is determined by a
descent algorithm.
5
1 / 8
0
B
1
In this scheme, /
8!
0 are computed by solving the standard quadratic optiB
mization problem (2). Our implementation, based on an interior point method, will not
be detailed here. Several SVM retraining are necessary, but they are faster than the usual
training since the algorithm is initialized appropriately with the solutions of the preceding
round.
For solving the minimization problem with respect to , we use a reduced conjugate gradient technique. The optimization problem was simplified by assuming that some of the other
variables are fixed. We tried several versions: 1) fixed; 2) Lagrange multipliers
fixed;
3) set of support vectors fixed. For the three versions, the optimal value of , or at least the
optimal value of the slack variables
can be obtained by solving a linear program, whose
optimum is computed directly (in a single iteration). We do not detail our first version here,
since the two last ones performed much better. The main steps of the two last versions are
sketched below.
1
8
B
B
1
8/ 0
?
<
1
1 > 5
Regarding problem (3), 1 is sub-optimal when
3.4 Sclaling parameters update
Starting from an initial solution
, our goal is to update by solving a
simple intermediate problem providing an improved solution to the global problem (3). We
first assume that the Lagrange multipliers
defining are not affected by updates, so
that is defined as
.
1
1
varies; nevertheless is guaranteed to
be an admissible solution. Hence, we minimize an upper bound of the original primal
cost which guarantees that any admissible update (providing a decrease of the cost) of the
intermediate problem will provide a decrease of the cost of the original problem.
The intermediate optimization problem is stated as follows:
? <? <
> >
)
<
? @
<
>
:
>
subject to
<
&
: &%
&%
!
:<
@
<B
<
:
<
$ 7!
<CB
< BB
798 B B
'
BB &
(4)
Solving this problem is still difficult since the cost is a complex non-linear function of
scale factors. Hence, as stated above, will be updated by a descent algorithm. The latter
requires the evaluation of the cost and its gradient with respect to . In particular, this
means that we should be able to compute
and
for any value of .
<
<
<
<
'
For given values of
and
,
is the solution of the following problem:
<
<
:
)
-
<
? @ *< B
: > 798
subject to >
<
!
whose dual formulation is
>
:<
?
>
< <
:< >
<
subject to
<
<
?
@
<B
BB
B B B
(5)
: >
"
!
@
<
<
BB
(6)
<0B
" >
<
< <
and its derivative with respect to are easily computed. Parameters
With ,
by a conjugate reduced gradient technique, i.e. a conjugate gradient
are then updated
algorithm ensuring that the set of constraints on are always verified.
This linear problem is solved directly by the following algorithm: 1) sort
in descending order for all positive examples on the one
and for all negative examples on the other side; 2) compute the pairwise sum of sorted
side
values; 3) set
for all positive and negative examples whose sum is positive.
3.5 Updating Lagrange multipliers
B B
' 8 < <
>
1. for support vectors of the first category 2
? @ <*B
<
: > 7 8 >
<
Assume now that only the support vectors remain fixed while optimizing . This assump
tion is used to derive a rule to update at
reasonable computing cost the Lagrange multipliers
together with by computing
. At
, the following holds [3]:
2. for support vectors of the second category (such that
(7)
?"<
)
.
From these equations, and the assumption that support vectors remain support vectors (and
that their category do not change) one derives a system of linear equations defining the
derivatives of
and with respect to [3]:
8
?
1. for support vectors of the first category
@ <B
?
@ <B
> $ 7 : > 7 #8
:
?<
2. for support vectors of the second category
'
(8)
?
>
:
obey the constraint
3. Finally, the system is completed by stating that the Lagrange multipliers should
:
?
:
>
? <
(9)
1
The value of
is updated from these equations, and the step size is limited to ensure that
for support vectors of the first category. Hence, in this version, is also an
admissible sub-optimal solution regarding problem (3).
4 Experiments
In the experiments reported below, we used (
for the constraint on (3). The scale parameters were optimized with the last version, where the set of support vectors is assumed
were chosen using the span bound [3].
to be fixed. Finally, the hyper-parameters
Although the value of the bound itself was not a faithful estimate of test error, the average
loss induced by using the minimizer of these bounds was quite small.
B
%
4.1 Toy experiment
In [9], Weston et al. compared two versions of their feature selection algorithm, to standard
SVMs and filter methods (i.e. preprocessing methods selecting features either based on
Pearson correlation coefficients, Fisher criterion score, or the Kolmogorov-Smirnov statistic). Their artificial data benchmarks provide a basis for comparing our approach with
their, which is based on the minimization of error bounds. Two types of distributions are
provided, whose detailed characteristics are not given here. In the linear problem, 6 dimensions out of 202 are relevant. In the nonlinear problem, two features out of 52 are relevant.
For each distribution, 30 experiments are conducted, and the average test recognition rate
measures the performance of each method.
For both problems, standard SVM achieve a 50% error rate in the considered range of
training set sizes. Our results are shown in Figure 1.
0.5
0.5
0.4
0.4
0.3
0.3
0.2
0.2
0.1
0.1
0
10
20
30
40
50
75
100
0
10
20
30
40
50
75
100
Figure 1: Results obtained on the benchmarks of [9]. Left: linear problem; right nonlinear
problem. The number of training examples is represented on the -axis, and the average
test error rate on the -axis.
>
Our test performances are qualitatively similar to the ones obtained by gradient descent on
the radius/margin bound in [9], which are only improved by the forward selection algorithm
minimizing the span bound. Note however that Weston et al. results are obtained after a
correct number of features was specified by the user, whereas the present results were
obtained fully automatically. Knowing the number of features that should be selected by
the algorithm is somewhat similar
to select the optimal value of parameter ( for each .
'%
problem, for
In the non-linear
training examples, an average of 26.5 features are
selected; for
8 , an average of 6.6 features are selected. These figures show that
although our feature selection scheme is effective, it should be more stringent: a smaller
variables
value of ( would
of problem. The two relevant
more appropriate for this type
selected in be
are
of cases for
and
, in for n=50, and in for
. For these two sample sizes, they
and second.
are even always ranked first
Regarding
training times, the optimization of required an average of over 100 times
more computing time than standard SVM fitting for the linear problem and 40 times for the
nonlinear problem. These increases scale less than linearly with the number of variables,
and are certainly yet to be improved.
4.2 Expression recognition
We also tested our algorithm on a more demanding task to test its ability to handle a large
number of features. The considered problem consists in recognizing the happiness expression among the five other facial expressions corresponding to universal emotions (disgust,
sadness, fear, anger, and surprise). The data sets are made of
8 gray level images of
frontal faces, with standardized positions of eyes, nose and mouth. The training set comprises
8 positive images, and negative ones. The test set is made of positive images
and negative ones.
the raw pixel representation of images, resulting in 4200 highly correlated feaWe used
tures. For this task, the accuracy of standard SVMs is 92.6% (11 test errors). The recognition rate is not significantly affected by our feature selection scheme (10 errors), but more
than 1300 pixels are considered to be completely irrelevant at the end of the iterative procedure (estimating required about 80 times more computing time than standard SVM).
This selection brings some important clues for building relevant attributes for the facial
recognition expression task.
Figure 2 represents the scaling factors , where black is zero and white represents the
highest value. We see that, according to the classifier, the relevant areas for recognizing the
happiness expression are mainly in the mouth area, especially on the mouth wrinkles, and
to a lesser extent in the white of the eyes (which detects open eyes) and the outer eyebrows.
On the right hand side of this figure, we displayed masked support faces, i.e. support faces
scaled by the expression mask. Although we lost many important features regarding the
identity of people, the expression is still visible on these faces. Areas irrelevant for the
recognition task (forehead, nose, and upper cheeks) have been erased or softened by the
expression mask.
5 Conclusion
We have introduced a method to perform automatic relevance determination and feature
selection in nonlinear SVMs. Our approach considers that the metric in input space defines
a set of parameters of the SVM classifier. The update of the scale factors is performed
by iteratively minimizing an approximation of the SVM cost. The latter is efficiently minimized with respect to slack variables when the metric varies. The approximation of the cost
function is tight enough to allow large update of the metric when necessary. Furthermore,
because at each step our algorithm guaranties the global cost to decrease, it is stable.
Figure 2: Left: expression mask of happiness provided by the scaling factors ; Right,
top row: the two positive masked support face; Right, bottom row: four negative masked
support faces.
Preliminary experimental results show that the method provides sensible results in a reasonable time, even in very high dimensional spaces, as illustrated on a facial expression
recognition task. In terms of test recognition rates, our method is comparable with [9, 3].
Further comparisons are still needed to demonstrate the practical merits of each paradigm.
%
Finally, it may also be beneficial to mix the two approaches: the method of Cristianini et al.
[4] could be used to determine and
. The resulting algorithm would differ from [9, 3],
since the relative relevance of each feature (as measured by ) would be estimated by
empirical risk minimization, instead of being driven by an estimate of generalization error.
%
References
[1] P. S. Bradley and O. L. Mangasarian. Feature selection via concave minimization and
support vector machines. In Proc. 15th International Conf. on Machine Learning,
pages 82?90. Morgan Kaufmann, San Francisco, CA, 1998.
[2] L. Breiman. Heuristics of instability and stabilization in model selection. The Annals
of Statistics, 24(6):2350?2383, 1996.
[3] O. Chapelle, V. Vapnik, O. Bousquet, and S. Mukherjee. Choosing multiple parameters
for support vector machines. Machine Learning, 46(1):131?159, 2002.
[4] N. Cristianini, C. Campbell, and J. Shawe-Taylor. Dynamically adapting kernels in
support vector machines. In M. S. Kearns, S. A. Solla, and D. A. Cohn, editors, Advances in Neural Information Processing Systems 11. MIT Press, 1999.
[5] T. Hastie, R. Tibshirani, and J. Friedman. The Elements of Statistical Learning: data
mining , inference, and prediction. Springer series in statistics. Springer, 2001.
[6] T. Jebara and T. Jaakkola. Feature selection and dualities in maximum entropy discrimination. In Uncertainity In Artificial Intellegence, 2000.
[7] R. M. Neal. Bayesian Learning for Neural Networks, volume 118 of Lecture Notes in
Statistics. Springer, 1996.
[8] V. N. Vapnik. The Nature of Statistical Learning Theory. Springer Series in Statistics.
Springer, 1995.
[9] J. Weston, S. Mukherjee, O. Chapelle, M. Pontil, T. Poggio, and V. Vapnik. Feature
selection for SVMs. In Advances in Neural Information Processing Systems 13. MIT
Press, 2000.
| 2156 |@word trial:3 illustrating:1 version:7 norm:3 smirnov:1 retraining:1 open:1 tried:1 initial:1 series:3 score:1 selecting:2 tuned:2 bradley:1 comparing:1 assigning:1 yet:1 visible:1 interpretable:1 update:7 discrimination:1 selected:4 egne:2 provides:2 five:1 become:1 consists:5 fitting:1 pairwise:1 mask:3 notably:1 indeed:1 multi:1 relying:1 utc:1 detects:1 automatically:2 encouraging:1 provided:4 estimating:1 minimizes:1 finding:2 transformation:1 guarantee:1 concave:1 universit:1 classifier:16 demonstrates:1 scaled:1 control:1 positive:7 local:1 insa:2 limit:1 might:1 black:1 umr:1 dynamically:1 sadness:1 limited:1 range:1 faithful:1 practical:1 lost:1 procedure:10 pontil:1 area:3 universal:1 empirical:5 significantly:1 wrinkle:1 adapting:1 suggest:1 interior:1 selection:22 risk:5 context:1 instability:1 descending:1 modifies:2 exposure:1 go:1 starting:2 convex:2 rule:1 handle:1 updated:3 annals:1 user:2 element:2 recognition:10 updating:1 mukherjee:2 bottom:1 role:1 solved:1 solla:1 decrease:3 highest:1 complexity:2 constrains:2 technologie:1 cristianini:3 solving:6 tight:1 basis:1 completely:1 easily:1 mization:1 represented:1 kolmogorov:1 fast:1 effective:1 pertaining:1 artificial:2 hyper:16 outcome:1 pearson:1 choosing:1 whose:5 quite:1 guaranty:1 solve:2 heuristic:1 say:1 favor:1 statistic:5 ability:1 itself:1 propose:1 fr:2 relevant:6 achieve:1 optimum:1 produce:2 derive:2 stating:1 measured:3 ard:1 differ:2 uncertainity:1 radius:2 correct:1 stephane:1 filter:1 attribute:1 vc:1 stabilization:1 stringent:1 proviso:1 generalization:10 preliminary:1 hold:2 considered:4 cb:2 mapping:5 smallest:1 purpose:1 estimation:1 proc:1 combinatorial:1 bridge:1 minimization:9 mit:2 gaussian:1 always:2 aim:1 avoid:3 breiman:1 jaakkola:1 heudiasyc:1 mainly:1 inference:1 cnrs:1 typically:1 kernelized:1 france:2 sketched:1 pixel:2 classification:2 among:3 dual:1 priori:1 rouvray:1 art:1 emotion:1 once:1 encouraged:1 represents:2 anger:1 minimized:2 ephane:1 optib:1 replaced:1 ab:1 attempt:1 friedman:1 highly:2 mining:1 evaluation:1 certainly:1 introduces:2 primal:1 encourage:2 necessary:2 poggio:1 facial:4 indexed:2 taylor:1 initialized:1 formalism:1 soft:2 cost:9 happiness:3 subset:1 masked:3 recognizing:2 successful:1 conducted:1 reported:1 varies:2 st:2 international:1 picking:2 together:1 conf:1 resort:1 derivative:2 toy:1 de:3 rouen:2 summarized:1 coefficient:1 performed:4 later:1 tion:1 sort:1 minimize:2 yves:2 accuracy:1 kaufmann:1 characteristic:2 efficiently:1 yield:1 spaced:1 bayesian:3 raw:1 psi:1 tunable:2 campbell:1 improved:3 formulation:1 furthermore:1 correlation:1 hand:1 cohn:1 nonlinear:5 defines:2 brings:1 gray:1 grows:1 building:1 unbiased:1 multiplier:5 hence:6 iteratively:3 neal:1 illustrated:1 white:2 attractive:1 round:1 during:1 width:1 criterion:5 generalized:1 presenting:1 demonstrate:1 enlargement:1 image:4 mangasarian:1 exponentially:1 volume:1 forehead:1 automatic:3 grid:2 canu:2 zeroing:1 shawe:1 chapelle:3 stable:3 optimizing:4 irrelevant:4 driven:1 postponed:1 morgan:1 minimum:1 somewhat:1 preceding:1 determine:1 paradigm:2 multiple:2 mix:1 faster:1 determination:3 controlled:1 ensuring:1 prediction:1 metric:11 iteration:1 kernel:9 achieved:1 whereas:1 appropriately:1 biased:1 subject:5 induced:1 regularly:1 effectiveness:1 seem:1 structural:1 constraining:2 intermediate:3 enough:1 hastie:1 bandwidth:2 lasso:1 regarding:4 lesser:1 knowing:1 intensive:1 expression:11 remark:1 detailed:2 amount:2 svms:7 category:6 reduced:2 estimated:1 tibshirani:1 affected:2 four:1 nevertheless:1 acknowledged:1 monitor:1 achieving:1 penalizing:1 verified:1 kept:1 sum:2 disgust:1 reasonable:2 cheek:1 decision:1 scaling:9 comparable:1 bound:10 followed:1 guaranteed:2 quadratic:1 adapted:1 constraint:8 bousquet:1 span:2 softened:1 according:2 conjugate:3 remain:2 smaller:1 beneficial:1 equation:3 slack:3 count:1 needed:1 letting:2 nose:2 merit:1 end:1 generalizes:1 apply:1 obey:1 appropriate:3 frequentist:2 original:2 standardized:1 top:1 ensure:1 completed:2 etienne:1 especially:1 added:1 usual:4 diagonal:3 gradient:5 outer:1 sensible:1 extent:1 unstable:1 considers:1 assuming:1 providing:2 minimizing:5 difficult:2 favorably:1 stated:2 negative:5 implementation:1 perform:2 upper:2 benchmark:2 descent:3 displayed:1 defining:5 jebara:1 introduced:1 required:2 specified:1 conclusive:1 optimized:4 unicity:1 address:1 able:1 suggested:1 usually:1 pattern:1 below:2 sparsity:3 eyebrow:1 program:1 including:1 mouth:3 demanding:2 ranked:1 scheme:5 eye:3 axis:2 alternated:1 determining:1 relative:1 loss:1 fully:1 lecture:1 tures:1 penalization:3 principle:1 editor:1 grandvalet:2 row:2 last:3 side:3 allow:3 face:6 sparse:2 benefit:1 dimension:2 author:1 qualitatively:1 adaptive:5 preprocessing:2 simplified:1 forward:1 made:2 clue:1 san:1 global:4 overfitting:2 assumed:1 francisco:1 continuous:1 iterative:2 hazardous:1 nature:1 ca:1 du:1 complex:2 diag:1 main:1 linearly:1 hyperparameters:1 sub:2 position:1 comprises:1 explicit:1 admissible:3 down:1 svm:21 derives:1 vapnik:3 compi:2 magnitude:2 margin:2 surprise:1 entropy:1 assump:1 lagrange:5 fear:1 springer:5 minimizer:1 weston:4 goal:1 sorted:1 identity:1 twofold:1 erased:1 fisher:1 hard:2 change:1 determined:2 kearns:1 duality:1 experimental:3 select:2 support:20 people:1 latter:3 relevance:6 frontal:1 tested:1 correlated:1 |
1,270 | 2,157 | Learning to Perceive Transparency from
the Statistics of Natural Scenes
Anat Levin
Assaf Zomet
Yair Weiss
School of Computer Science and Engineering
The Hebrew University of Jerusalem
91904 Jerusalem, Israel
{alevin,zomet,yweiss}@cs.huji.ac.il
Abstract
Certain simple images are known to trigger a percept of transparency: the input image I is perceived as the sum of two images
I(x, y) = I1 (x, y) + I2 (x, y). This percept is puzzling. First, why
do we choose the ?more complicated? description with two images
rather than the ?simpler? explanation I(x, y) = I1 (x, y) + 0 ? Second, given the infinite number of ways to express I as a sum of two
images, how do we compute the ?best? decomposition ?
Here we suggest that transparency is the rational percept of a system that is adapted to the statistics of natural scenes. We present
a probabilistic model of images based on the qualitative statistics
of derivative filters and ?corner detectors? in natural scenes and
use this model to find the most probable decomposition of a novel
image. The optimization is performed using loopy belief propagation. We show that our model computes perceptually ?correct?
decompositions on synthetic images and discuss its application to
real images.
1
Introduction
Figure 1a shows a simple image that evokes the percept of transparency. The
image is typically perceived as a superposition of two layers: either a light square
with a dark semitransparent square in front of it or a dark square with a light
semitransparent square in front of it.
Mathematically, our visual system is taking a single image I(x, y) and representing
as the sum of two images:
I1 (x, y) + I2 (x, y) = I(x, y)
(1)
When phrased this way, the decomposition is surprising. There are obviously an
infinite number of solutions to equation 1, how does our visual system choose one?
Why doesn?t our visual system prefer the ?simplest? explanation I(x, y) = I 1 (x, y)+
0?
a
b
Figure 1: a. A simple image that evokes the percept of transparency. b. A simple
image that does not evoke the percept of transparency.
Figure 1b shows a similar image that does not evoke the percept of transparency.
Here again there are an infinite number of solutions to equation 1 but our visual
system prefers the single layer solution.
Studies of the conditions for the percept of transparency go back to the very first research on visual perception (see [1] and references within). Research of this type has
made great progress in understanding the types of junctions and their effects (e.g.
X junctions of a certain type trigger transparency, T junctions do not). However,
it is not clear how to apply these rules to an arbitrary image.
In this paper we take a simple Bayesian approach. While equation 1 has an infinite
number of possible solutions, if we have prior probabilities P (I1 (x, y)), P (I2 (x, y))
then some of these solutions will be more probable than others. We use the statistics
of natural images to define simple priors and finally use loopy belief propagation
to find the most probable decomposition. We show that while the model knows
nothing about ?T junctions? or ?X junctions?, it can generate perceptually correct
decompositions from a single image.
2
Statistics of natural images
A remarkably robust property of natural images that has received much attention
lately is the fact that when derivative filters are applied to natural images, the filter
outputs tend to be sparse [5, 7]. Figure 2 illustrates this fact: the histogram of the
horizontal derivative filter is peaked at zero and fall off much faster than a Gaussian.
Similar histograms are observed for vertical derivative filters and for the gradient
magnitude: |?I|.
There are many ways to describe the non Gaussian nature of this distribution
(e.g. high kurtosis, heavy tails). Figure 2b illustrates the observation made by
Mallat [4] and Simoncelli [8]: that the distribution is similar to an exponential
density with exponent
less than 1. We show the log probability for densities of the
?
form p(x) ? e?x . We assume x ? [0, 100] and plot the log probabilities so that
they agree on p(0), p(100). There is a qualitative difference between distributions
for which ? > 1 (when the log probability is convex) and those for which ? < 1
(when it becomes concave). As figure 2d shows, the natural statistics for derivative
deriv filter
corner operator
5
5
2.5
x 10
5
x 10
4.5
4
2
3.5
3
1.5
2.5
2
1
1.5
1
0.5
0.5
0
?150
0
?0.5
?100
?50
0
a
150
200
0
0.5
1
1.5
2
2.5
7
250
x 10
e
14
0
12
?2
10
?4
2
Gaussian:?x
Laplacian: ?x
?0.4
logprob
100
c
0
?0.2
50
1/2
?X
?6
8
1/4
?X
?8
?0.6
6
?10
4
?0.8
?12
2
?1
0
50
100
x
b
150
?14
0
0
50
100
150
200
0
0.5
1
1.5
d
2
2.5
7
250
x 10
f
Figure 2: a. A natural image. c Histogram of filter outputs. e Histogram of corner
detector outputs. d,e log histograms.
?
filters has the qualitative nature of a distribution e?x with ? < 1.
In [9] the sparsity of derivative filters was used to decompose an image sequence as
a sum of two image sequences. Will this prior be sufficient for a single frame ? Note
that decomposing the image in figure 1a into two layers does not change the output
of derivative filters: exactly the same derivatives exist in the single layer solution
as in the two layer solution. Thus we cannot appeal to the marginal histogram of
derivative filters to explain the percept of transparency in this image.
There are two ways to go beyond marginal histograms of derivative filters. We
can either look at joint statistics of derivative filters at different locations or orientations [6] or look at marginal statistics of more complicated feature detectors
(e.g. [11]).
We looked at the marginal statistics of a ?corner detector?. The output of the
?corner detector? at a given location x0 , y0 is defined as:
?
?
X
Ix2 (x, y)
Ix (x, y)Iy (x, y)
c(x0 , y0 ) = det(
w(x, y)
)
(2)
Ix (x, y)Iy (x, y)
Iy2 (x, y)
where w(x, y) is a small Gaussian window around x0 , y0 and Ix , Iy are the derivatives
of the image.
Figures 2e,f show the histogram of this corner operator on a typical natural
image.
?
Again, note that it has the qualitative statistic of a distribution e?x for ? < 1.
To get a more quantitative description of the statistics
we used maximum likelihood
?
to fit a distribution of the form P (x) = Z1 e?ax to gradient magnitudes and corner
detector histograms in a number of images. We found that the histograms shown
in figure 2 are typical: for both gradients and corner detectors the exponent was
less than 1 and the exponent for the corner detector was smaller than that of the
gradients. Typical exponents were 0.7 for the derivative filter and 0.25 for the corner
detector. The scaling parameter a of the corner detector was typically larger than
that of the gradient magnitude.
3
Simple prior predicts transparency
Motivated by the qualitative statistics observed in natural images we now define a
probability distribution over images. We define the log probability of an image by
means of a probability over its gradients:
X?
?
log P (Ix , Iy ) = ? log Z ?
|?I(x, y)|? + ?c(x, y)?
(3)
x,y
with ? = 0.7, ? = 0.25. The parameter ? was determined by the ratio of the scaling
parameters in the corner and gradient distributions.
Given a candidate decomposition of an image I into I1 and I2 = I ? I1 we define
the log probability of the decomposition as the sum of the log probabilities of the
gradients of I1 and I2 . Of course this is only an approximation: we are ignoring
dependencies between the gradients across space and orientation. Although this is
a weak prior, one can ask: is this enough to predict transparency? That is, is the
most probable interpretation of figure 1a one with two layers and the most probable
decomposition of figure 1b one with a single layer?
Answering this question requires finding the global maximum of equation 3. To
gain some intuition we calculated the log probability of a one dimensional family
of solutions. We defined s(x, y) the image of a single white square in the same
location as the bottom right square in figure 1a,b. We considered decompositions
of the form I1 = ?s(x, y),I2 = I ? I1 and evaluated the log probability for values of
? between ?1 and 2.
Figure 3a shows the result for figure 1a. The most probable decomposition is the
one that agrees with the percept: ? = 1 one layer for the white square and another
for the gray square. Figure 3b shows the result for figure 1b. The most probable
decomposition again agrees with the percept: ? = 0 so that one layer is zero and
the second contains the full image.
3.1
The importance of being non Gaussian
Equation 3 can be verbally described as preferring decompositions where the total
edge and corner detector magnitudes are minimal. Would any cost function that
has this preference give the same result?
Figure 3c shows the result with ? = ? = 2 for the transparency figure (figure 1a).
This would be the optimal interpretation if the marginal histograms of edge and
corner detectors were Gaussian. Now the optimal interpretation indeed contains
two layers but they are not the ones that humans perceive. Thus the non Gaussian
nature of the histograms is crucial for getting the transparency percept. Similar
?non perceptual? decompositions are obtained with other values of ?, ? > 1.
We can get some intuition for the importance of having exponents smaller than
1 from the following observation which considers the analog of the transparency
problem with scalars. We wish to solve the equation a + b = 1 and we have a prior
over positive scalars of the form P (x).
Observation: The MAP solution to the scalar transparency problem is obtained
with a = 1, b = 0 or a = 0, b = 1 if and only if log P (x) is concave.
The proof follows directly from the definition of concavity.
160
800
I1=?
I=
I1= ?
-log(prob)
600
-log(prob)
120
160
100
I1= ?
I=
700
180
-log(prob)
140
200
I=
500
400
140
300
80
120
60
-1
100
-1
0
?
1
2
200
?
0
a
1
2
100
-1
b
?
0
1
2
c
Figure 3: a-b. negative log probability (equation 3) for a sequence of decompositions of figure 1a,b respectively. The first layer is always a single square with
contrast ? and the second layer is shown in the insets. c. negative log probability
(equation 3) for a sequence of decompositions of figure 1a with ? = ? = 2.
4
Optimization using loopy BP
Finding the most likely decomposition requires a highly nonlinear optimization. We
chose to discretize the problem and use max-product loopy belief propagation to find
the optimum. We defined a graphical model in which every node gi corresponded to
a discretization of the gradient of one layer I1 at that location gi = (gix , giy )T . For
every value of gi we defined fi which represents the gradient of the second layer at
that location: fi = (Ix , Iy )T ? gi . Thus the two gradients fields {gi }, {fi } represent
a valid decomposition of the input image I.
The joint probability is given by:
Y
1 Y
P (g) =
?i (gi )
?ijkl (gi , gj , gk , gl )
Z i
(4)
<ijkl>
where < ijkl > refers to four adjacent pixels that form a 2x2 local square.
The local potential ?i (gi ) is based on the histograms of derivative filters:
?i (gi ) = e(?|g|
?
?|f |? )/T
(5)
where T is an arbitrary system ?temperature?.
The fourway potential: ?ijkl (gi , gj , gk , gl ) is based on the histogram of the corner
operator:
?ijkl (gi , gj , gk , gl ) = e??/T (det(gi gi
T
T
+gj gjT +gk gk
+gl glT )? +det(fi fiT +fj fjT +fk fkT +fl flT )? )
(6)
To enforce integrability of the gradient fields the fourway potential is set to zero
when gi , gj , gk , gl violate the integrability constraint (cf. [3]).
The graphical model defined by equation 4 has many loops. Nevertheless motivated
by the recent results on similar graphs [2, 3] we ran the max-product belief propagation algorithm on it. The max-product algorithm finds a gradient field {g i } that
is a local maximum of equation 4 with respect to a large neighbourhood [10]. This
gradient field also defines the complementary gradient field {fi } and finally we integrate the two gradient fields to find the two layers. Since equation 4 is completely
symmetric in {f } and {g} we break the symmetry by requiring that the gradient
in a single location gi0 belong to layer 1.
In order to run BP we need to somehow discretize the space of possible gradients
at each pixel. Similar to the approach taken in [2] we use the local potentials to
Input I
Output I1
Output I2
Figure 4: Output of the algorithm on synthetic images. The algorithm effectively
searches over an exponentially large number of possible decompositions and chooses
decompositions that agree with the percept.
sample a small number of candidate gradients at each pixel. Since the local potential
penalizes non zero gradients, the most probable candidates are gi = (Ix , Iy ) and
gi = (0, 0). We also added two more candidates at each pixel gi = (Ix , 0) and
gi = (0, Iy ). With this discretization there are still an exponential number of
possible decompositions of the image. We have found that the results are unchanged
when more candidates are introduced at each pixel.
Figure 4 shows the output of the algorithm on the two images in figure 1. An
animation that illustrates the dynamics of BP on these images is available at
www.cs.huji.ac.il/ ?yweiss. Note that the algorithm is essentially searching exponentially many decompositions of the input images and knows nothing about ?X
junctions? or ?T junctions? or squares. Yet it finds the decompositions that are
consistent with the human percept.
Will our simple prior also allow us to decompose a sum of two real images ? We
first tried a one dimensional family of solutions as in figure 3. We found that for
real images that have very little texture (e.g. figure 5b) the maximal probability
solution is indeed obtained at the perceptually correct solution. However, nearly
any other image that we tried had some texture and on such images the model failed
(e.g. 5a). When there is texture in both layers, the model always prefers a one layer
decomposition: the input image plus a zero image. To understand this failure,
recall that the model prefers decompositions that have few corners and few edges.
According to the simple ?edge? and ?corner? operators that we have used, real
images have edges and corners at nearly every pixel so the two layer decomposition
has twice as many edges and corners as the one layer decomposition. To decompose
general real images we need to use more sophisticated features to define our prior.
Even for images with little texture standard belief propagation with synchronous
a
b
c
d
Figure 5: When we sum two arbitrary images (e.g. in a.) the model usually prefers
the one layer solution. This is because of the texture that results in gradients and
corners at every pixel. For real images that are relatively texture free (e.g. in b.)
the model does prefer splitting into two layers (c. and d.)
updates did not converge. Significant manual tweaking was required to get BP to
converge. First, we manually divided the input image into smaller patches and ran
BP separately on each patch. Second, to minimize discretization artifacts we used
a different number of gradient candidates at each pixel and always included the
gradients of the original images in the list of candidates at that pixel. Third, to
avoid giving too much weight to corners and edges in textured regions, we increased
the temperature at pixels where the gradient magnitude was not a local maximum.
The results are shown at the bottom of 5. In preliminary experiments we have found
that similar results can be obtained with far less tweaking when we use generalized
belief propagation to do the optimization.
5
Discussion
The percept of transparency is a paradigmatic example of the ill-posedness of vision:
the number of equations is half the number of unknowns. Nevertheless our visual
systems reliably and effectively compute a decomposition of a single image into
two images. In this paper we have argued that this perceptual decomposition may
correspond to the most probable decomposition using a simple prior over images
derived from natural scene statistics.
We were surprised with the mileage we got out of the very simple prior we used: even
though it only looks at two operators (gradients, and cornerness) it can generate
surprisingly powerful predictions. However, our experiments with real images show
that this simple prior is not powerful enough. In future work we would like to
add additional features. One way to do this is by defining features that look for
?texture edges? and ?texture corners? and measuring their statistics in real images.
A second way to approach this is to use a full exponential family maximum likelihood
algorithm (e.g. [11]) that automatically learned which operators to look at as well
as the weights on the histograms.
References
[1] E.H. Adelson. Lightness perception and lightness illusions. In M. Gazzaniga,
editor, The new cognitive neurosciences, 2000.
[2] W.T. Freeman and E.C. Pasztor. Learning to estimate scenes from images.
In M.S. Kearns, S.A. Solla, and D.A. Cohn, editors, Adv. Neural Information
Processing Systems 11. MIT Press, 1999.
[3] B.J. Frey, R. Koetter, and N. Petrovic. Very loopy belief propagation for
unwrapping phase images. In Adv. Neural Information Processing Systems 14.
2001.
[4] S. Mallat. A theory for multiresolution signal decomposition : the wavelet
representation. IEEE Trans. PAMI, 11:674?693, 1989.
[5] B.A. Olshausen and D. J. Field. Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature, 381:607?608, 1996.
[6] J. Portilla and E. P. Simoncelli. A parametric texture model based on joint
statistics of complex wavelet coefficients. Int?l J. Comput. Vision, 40(1):49?71,
2000.
[7] E.P. Simoncelli. Statistical models for images:compression restoration and synthesis. In Proc Asilomar Conference on Signals, Systems and Computers, pages
673?678, 1997.
[8] E.P. Simoncelli. Bayesian denoising of visual images in the wavelet domain. In
P Mller and B Vidakovic, editors, Wavelet based models, 1999.
[9] Y. Weiss. Deriving intrinsic images from image sequences. In Proc. Intl. Conf.
Computer Vision, pages 68?75. 2001.
[10] Y. Weiss and W.T. Freeman. On the optimality of solutions of the maxproduct belief propagation algorithm in arbitrary graphs. IEEE Transactions
on Information Theory, 47(2):723?735, 2001.
[11] Song Chun Zhu, Zing Nian Wu, and David Mumford. Minimax entropy principle and its application to texture modeling. Neural Computation, 9(8):1627?
1660, 1997.
| 2157 |@word semitransparent:2 compression:1 tried:2 decomposition:31 contains:2 discretization:3 surprising:1 yet:1 koetter:1 nian:1 plot:1 update:1 half:1 node:1 location:6 preference:1 simpler:1 surprised:1 qualitative:5 assaf:1 x0:3 indeed:2 freeman:2 automatically:1 little:2 window:1 becomes:1 israel:1 finding:2 quantitative:1 every:4 concave:2 exactly:1 positive:1 engineering:1 local:6 frey:1 pami:1 chose:1 plus:1 twice:1 illusion:1 got:1 refers:1 tweaking:2 suggest:1 get:3 cannot:1 operator:6 www:1 map:1 jerusalem:2 go:2 attention:1 convex:1 splitting:1 perceive:2 rule:1 deriving:1 searching:1 trigger:2 mallat:2 predicts:1 observed:2 bottom:2 region:1 adv:2 solla:1 ran:2 intuition:2 dynamic:1 textured:1 completely:1 joint:3 describe:1 mileage:1 corresponded:1 larger:1 solve:1 statistic:15 gi:18 emergence:1 obviously:1 sequence:5 kurtosis:1 product:3 maximal:1 loop:1 multiresolution:1 description:2 getting:1 optimum:1 intl:1 ac:2 school:1 received:1 progress:1 c:2 correct:3 filter:15 human:2 argued:1 yweiss:2 decompose:3 preliminary:1 probable:9 mathematically:1 around:1 considered:1 great:1 predict:1 mller:1 perceived:2 proc:2 superposition:1 agrees:2 mit:1 gaussian:7 always:3 rather:1 avoid:1 ax:1 derived:1 likelihood:2 integrability:2 contrast:1 typically:2 i1:14 pixel:10 orientation:2 ill:1 exponent:5 marginal:5 field:8 having:1 manually:1 represents:1 look:5 adelson:1 nearly:2 peaked:1 future:1 others:1 few:2 phase:1 highly:1 light:2 edge:8 penalizes:1 minimal:1 increased:1 modeling:1 measuring:1 restoration:1 loopy:5 cost:1 levin:1 front:2 too:1 dependency:1 synthetic:2 chooses:1 petrovic:1 density:2 gix:1 huji:2 preferring:1 probabilistic:1 off:1 synthesis:1 iy:7 again:3 choose:2 gjt:1 corner:22 cognitive:1 conf:1 derivative:14 potential:5 coefficient:1 int:1 performed:1 break:1 complicated:2 minimize:1 il:2 square:11 percept:15 correspond:1 weak:1 bayesian:2 ijkl:5 ix2:1 detector:12 explain:1 manual:1 definition:1 failure:1 proof:1 rational:1 gain:1 ask:1 recall:1 sophisticated:1 back:1 wei:3 evaluated:1 though:1 horizontal:1 cohn:1 nonlinear:1 propagation:8 somehow:1 defines:1 artifact:1 gray:1 olshausen:1 effect:1 requiring:1 alevin:1 symmetric:1 i2:7 white:2 adjacent:1 generalized:1 temperature:2 fj:1 image:67 novel:1 fi:5 exponentially:2 tail:1 interpretation:3 analog:1 belong:1 significant:1 fk:1 had:1 gj:5 add:1 recent:1 certain:2 additional:1 converge:2 paradigmatic:1 signal:2 full:2 simoncelli:4 violate:1 transparency:17 faster:1 divided:1 laplacian:1 prediction:1 essentially:1 vision:3 histogram:15 represent:1 cell:1 remarkably:1 separately:1 crucial:1 tend:1 enough:2 fit:2 det:3 synchronous:1 motivated:2 song:1 prefers:4 clear:1 dark:2 simplest:1 generate:2 exist:1 neuroscience:1 express:1 four:1 nevertheless:2 graph:2 sum:7 run:1 prob:3 powerful:2 evokes:2 family:3 wu:1 patch:2 prefer:2 scaling:2 layer:22 fl:1 adapted:1 constraint:1 bp:5 scene:5 x2:1 phrased:1 optimality:1 relatively:1 according:1 smaller:3 across:1 y0:3 taken:1 asilomar:1 equation:12 agree:2 discus:1 know:2 junction:7 decomposing:1 available:1 apply:1 enforce:1 neighbourhood:1 yair:1 original:1 cf:1 graphical:2 giving:1 unchanged:1 question:1 added:1 looked:1 mumford:1 receptive:1 parametric:1 gradient:26 considers:1 code:1 ratio:1 hebrew:1 gk:6 negative:2 reliably:1 unknown:1 fjt:1 discretize:2 vertical:1 observation:3 pasztor:1 defining:1 frame:1 portilla:1 arbitrary:4 posedness:1 introduced:1 david:1 required:1 z1:1 learned:1 trans:1 beyond:1 gazzaniga:1 usually:1 perception:2 sparsity:1 max:3 explanation:2 belief:8 natural:13 zhu:1 representing:1 minimax:1 lightness:2 lately:1 prior:11 understanding:1 integrate:1 sufficient:1 consistent:1 principle:1 editor:3 gi0:1 unwrapping:1 heavy:1 course:1 gl:5 surprisingly:1 free:1 allow:1 understand:1 fall:1 taking:1 sparse:2 calculated:1 valid:1 computes:1 doesn:1 concavity:1 made:2 zing:1 far:1 transaction:1 evoke:2 global:1 search:1 why:2 nature:4 robust:1 ignoring:1 symmetry:1 complex:1 domain:1 did:1 animation:1 nothing:2 complementary:1 wish:1 exponential:3 comput:1 candidate:7 answering:1 perceptual:2 third:1 anat:1 ix:7 wavelet:4 inset:1 appeal:1 list:1 deriv:1 flt:1 chun:1 intrinsic:1 effectively:2 importance:2 texture:10 magnitude:5 perceptually:3 illustrates:3 entropy:1 likely:1 visual:7 failed:1 verbally:1 fkt:1 scalar:3 change:1 included:1 determined:1 infinite:4 typical:3 zomet:2 denoising:1 kearns:1 total:1 maxproduct:1 puzzling:1 |
1,271 | 2,158 | ynamic
Causal Learning
Thomas L. Griffiths
David Danks
Institute for Human & Machine Cognition
Department of Psychology
University of West Florida
Stanford University
Stanford, CA 94305-2130
Pensacola, FL 32501
[email protected]
[email protected]
Joshua B. Tenenbaum
Department of Brain & Cognitive Sciences
~AIT
Cambridge, MA 02139
[email protected]
Abstract
Current psychological theories of human causal learning and
judgment focus primarily on long-run predictions: two by
estimating parameters of a causal Bayes nets (though for different
parameterizations), and a third through structural learning. This
paper focuses on people's short-run behavior by examining
dynamical versions of these three theories, and comparing their
predictions to a real-world dataset.
1
Introduction
Currently active quantitative models of human causal judgment for single (and
sometimes multiple) causes include conditional j}JJ [8], power PC [1], and Bayesian
network structure learning [4], [9]. All of these theories have some normative
justification, and all can be understood rationally in terms of learning causal Bayes
nets. The first two theories assume a parameterization for a Bayes net, and then
perform maximum likelihood parameter estimation. Each has been the target of
numerous psychological studies (both confirming and disconfirming) over the past
ten years. The third theory uses a Bayesian structural score, representing the log
likelihood ratio in favor of the existence of a connection between the potential cause
and effect pair. Recent work found that this structural score gave a generally good
account, and fit data that could be fit by neither of the other two models [9].
To date, all of these models have addressed only the static case, in which judgments
are made after observing all of the data (either sequentially or in summary format).
Learning in the real world, however, also involves dynamic tasks, in which
judgments are made after each trial (or small number). Experiments on dynamic
tasks, and theories that model human behavior in them, have received surprisingly
little attention in the psychological community. In this paper, we explore dynamical
variants of each of the above learning models, and compare their results to a real
data set (from [7]). We focus only on the case of one potential cause, due to space
and theoretical constraints, and a lack of experimental data for the multivariate case.
2
Real-World Data
In the experiment on which we focus in this paper [7], people's stepwise acquisition
curves were measured by asking people to determine whether camouflage makes a
tank more or less likely to be destroyed. Subjects observed a sequence of cases in
which the tank was either camouflaged or not, and destroyed or not. They were
asked after every five cases to judge the causal strength of the- camouflage on a
[-100, +100] scale, where -100 and +100 respectively correspond to the potential
cause always preventing or producing the effect. The learning curves, constructed
from average strength ratings, were:
50
Positive contingent
High P(E) non-contingent
Low P(E) non-contingent
Negative contingent
Me
an
jud
gm
ent
-50
10
15
20
25
30
35
40
Trials
Figure 1: Example of learning curves
In this paper, we focus on qualitative features of the learning curves. These learning
curves can be divided on the basis of the actual contingencies in the experimental
condition. There were two contingent conditions: a positive condition in which
peE I C) = .75 (the probability of the effect given the cause) and peE I -,C) = .25,
and a negative condition where the opposite was true. There were also two noncontingent conditions, one in which peE) = .75 and one in which peE) = .25,
irrespective of the presence or absence of the causal variable. We refer to the former
non-contingent condition as having a high peE), and the latter as having a low peE).
There are two salient, qualitative features of the acquisition curves:
3
3.1
1.
For contingent cases, the strength rating does not immediately reach the
final judgment, but rather converges to it slowly; and
2.
For non-contingent cases, there is an initial non-zero strength rating when
the probability of the effect, peE), is high, followed by convergence to zero.
Parameter Estimation Theories
Conditional
~p
The conditional f1P theory predicts that the causal strength rating for a particular
factor will be (proportional to) the conditional contrast for that factor [5], [8]. The
general form of the conditional contrast for a particular potential cause is given by:
f1P C.{X} = peE I C & X) - peE I -,C & X), where X ranges over the possible states of
the other potential causes. So, for example, if we have two potential causes, C 1 and
C2 , then there are two conditional contrasts for C 1 : f1P C l.{C2} = peE I C1 & C2 ) peE I -'C 1 & C2 ) and f1P C l.{-.C2} = peE I C1 & -,C2 ) - peE I-'C1 & -,C2 ). Depending
on the probability distribution, some conditional contrasts for a potential cause may
be undefined, and the defined contrasts for a particular variable may not agree. The
conditional I1P theory only makes predictions about a potential cause when the
underlying probability distribution is "well-behaved": at least one of the conditional
contrasts for the factor is defined, and all of the defined conditional contrasts for the
factor are equal. For a single cause-effect relationship, calculation of the J1P value is
a maximum likelihood parameter estimator assuming that the cause and the
background combine linearly to predict the effect [9J.
Any long-run learning model can model sequential data by being applied to all of
the data observed up to a particular point. That is, after observing n datapoints, one
simply applies the model, regardless of whether n is "the long-run." The behavior of
such a strategy for the conditional ~p theory is shown in Figure 2 (a), and clearly
fails to model accurately the above on-line learning curves. There is no gradual
convergence to asymptote in the contingent cases, nor is there differential behavior
in the non-contingent cases.
An alternative dynamical model is the Rescorla-Wagner model [6J, which has
essentially the same form as the well-known delta rule used for training simple
neural networks. The R-W model has been shown to converge to the conditionall1P
value in exactly the situations in which the I1P theory makes a prediction [2J. The
R-W model follows a similar statistical logic as the I1P theory: J1P gives the
maximum likelihood estimates in closed-form, and the R-W model essentially
implements gradient ascent on the log-likelihood surface, as the delta rule has been
shown to do. The R-W model produces' learning curves that qualitatively fit the
learning curves in Figure 1, but suffers from other serious flaws. For example,
suppose a subject is presented with trials of A, C, and E, followed by trials with only
A and E. In such a task, called backwards blocking, the R-W model predicts that C
should be viewed as moderately causal, but human subjects rate C as non-causal.
In the augmented R-W model [10J causal strength estimates (denoted by Vi, and
assumed to start at zero) change after each observed case. Assuming that b(.x) = 1 if
X occurs on a particular trial, and 0 otherwise, then strength estimates change by the
following equation:
aiO and ail are rate parameters (saliences) applied when Ci is present and absent,
respectively, and Po and PI are the rate parameters when E is present and absent,
respectively. By updating the causal strengths of absent potential causes, this model
is able to explain many of the phenomena that escape the normal R-W model, such
as backwards blocking.
Although the augmented R-W model does not always have the same asymptotic
behavior as the regular R-W model, it does have the same asymptotic behavior in
exactly those situations in which the conditional J1P theory makes a prediction
(under typical assumptions: aiO = -ail, Po = PI, and A = 1) [2]. To determine whether
the augmented R-W model also captures the qualitative features of people's
dynamic learning, we performed a simulation in which 1000 simulated individuals
were shown randomly ordered cases that matched the probability distributions used
in [7]. The model parameter values were A = 1.0, Q{)o = 0.4, alO = 0.7, au = -0.2, Po
= PI = 0.5, with two learned parameters: Vo for the always present background cause
Co, and VI for the potential cause C I . The mean values of VI, multiplied by 100 to
match scale with Figure 1, are shown in Figure 2 (b).
(b)
(a)
50
-50
5
10
15
20
25
30
35
40
10
15
20
25
30
35
40
5
10
15
20
25
30
35
40
5
10
15
20
25
30
35
40
50
50
-50
5
(e)
5
(d)
(c)
50~
,
=:.::=:=t
10
15
20
25
30
35
40
_5~~~~
5
10
15
20
25
30
35
40
-50
Figure 2: Modeling results. (a) is the maximum-likelihood estimate of fJJ , (b) is the
augmented R-W model, (c) is the maximum-likelihood estimate of causal power, (d)
is the analogue of augmented R-W model for causal power, (e) shows the Bayesian
strength estimate with a uniform prior on all parameters, and (f) does likewise with
a beta(I,5) prior on Va. The line-markers follow the conventions of Figure 1.
Variations in A only change the response scale. Higher values of lXoo (the salience of
the background) shift downward all early values of the learning curves, but do not
affect the asymptotic values. The initial non-zero values for the non-contingent
cases is proportional in size to (alO + al r), and so if the absence of the cause is more
salient than the presence, the initial non-zero value will actually be negative.
Raising the fJ values increases the speed of convergence to asymptote, and the
absolute values of the contingent asymptotes decrease in proportion to (fJo - fJI).
For the chosen parameter values, the learning curves for the contingent cases both
gradually curve towards an asymptote, and in the non-contingent, high peE) case,
there is an initial non-zero rating. Despite this qualitative fit and its computational
simplicity, the augmented R-W model does not have a strong rational motivation. Its
only rational justification is that it is a consistent estimator of fJJ: in the limit of
infinite data, it converges to fJJ under the same circumstances that the regular (and
well-motivated) R-W model does. But it does not seem to have any of the other
properties of a good statistical estimator: it is not unbiased, nor does it seem to be a
maximum likelihood or gradient-ascent-on-log-Iikelihood algorithm (indeed,
sometimes it appears to descend in likelihood). This raises the question of whether
there might be an alternative dynamical model of causal learning that produces the
appropriate learning curves but is also a principled, rational statistical estimator.
3.2
Power PC
In Cheng's power PC theory [1], causal strength estimates are predicted to be
(proportional to) perceived causal power: the (unobserved) probability that the
potential cause, in the absence of all other causes, will produce the effect. Although
causal power cannot be directly observed, it can be estimated from observed
statistics given some assumptions. The power PC theory predicts that, when the
assumptions are believed to be satisfied, causal power for (potentially) generative or
preventive 'causes will be estimated by the following equations:
?
p _
G eneratIve:
M
C
C-1-P(EI-,C)
Preventive: p =
e
- Me
p(EI-'C)
Because the power PC theory focuses on the long-run, one can easily d'etermine
which equation to use: simply wait until asymptote, determine J1P c , and then divide
by the appropriate factor. Similar equations can also be given for interactive causes.
Note that although the preventive causal power equation yields a positive number,
we should expect people to report a negative rating for preventive causes.
As with the t:JJ theory, the power PC theory can, in the case of a single cause-effect
pair, also be seen as a maximum li).<elihood estimator for the strength parameter of a
causal Bayes net, though one with a different parameterization than for conditional
t:JJ. Generative causes and the background interact to produce the effect as though
they were a noisy-OR gate. Preventive causes combine with them as a noisy-ANDNOT gate. Therefore, if the G/s are generative causes and lj's are preventive causes,
the theory predicts: P(E) =
I} (1- Ijl1- If (1- G,)]-
As for conditional J1P, simply applying the power PC equations to the sufficient
statistics for observed sequential data does not produce appropriate learning curves.
There is no gradual convergence in the contingent cases, and there is no initial
difference in the non-contingent cases. This behavior is shown in Figure 2 (c).
Instead, we suggest using an analogue of the augmented R-W model, which uses the
above noisy-ORlAND-NOT prediction instead of the linear prediction implicit in
the augmented R-W model. Specifically, we define the following algorithm (with all
parameters as defined before), using the notational device that the C/ s are
preventive and the Cj ' s are generative:
Unlike the R-W and augmented R-W models, there is no known characterization of
the long-run behavior of this iterative algorithm. However, we can readily determine
(using the equilibrium technique of [2]) the asymptotic Vi values for' one potential
cause (and a single, always present, generative background cause). If we make the
same simplifying assumptions as in Section 3.1, then this algorithm asymptotically
computes the causal power for C, regardless of whether C is generative or
preventive. We conjecture that this algorithm also computes the causal power for
multiple potential causes.
This iterative algorithm can only be applied if one knows whether each potential
cause is potentially generative or preventive. Furthermore, we cannot determine
directionality by the strategy of the power PC theory, as we do not necessarily have
the correct t:JJ sign during the short run. However, changing the classification of Ci
from generative to preventive (or vice versa) requires only removing from (adding
to) the estimate (i) the Vi term; and (ii) all terms in which Vi was the only generative
factor. Hence, we conjecture that this algorithm can be augmented to account for
reclassification of potential causes after learning has begun.
To simulate this dynamical version of the power PC theory, we used the same setup
as in Section 3 .1 (and multiplied preventive causal power ratings by -1 to properly
scale them). The parameters for this run were: A = 1.0, lXoo = 0.1, al0 = 0.5,
all = -0.4,/30 = /31 = 0.9, and the results are shown in Figure 2 (d). Parameter
variations have the same effects as for the augmented R-W model, except that
increasing lXoo reduces the size of the initial non-zero values in the non-contingent
conditions (instead of all conditions), and absolute values of the asymptotes in all
conditions are shifted by an amount proportional to (/30 - /31),
This dynamical theory produces the right sort of learning curves for these parameter
values, and is also a consistent estimator (converging to the power PC estimate in
the limit of infinite data). But as with the augmented R-W model, there is no
rational motivation for choosing this dynamic estimator: it is not unbiased, nor
maximum likelihood, nor an implementation of gradient ascent in log-likelihood.
The theory's main (and arguably only) advantage over the augmented R-W model is
that it converges to a quantity that is more typically what subj ects estimate in longrun experiments. But it is still not what we desire from a principled dynamic model.
4
Bayesian structure learning
The learning algorithms considered thus far are based upon the idea that human
causal judgments reflect the estimated value of a strength parameter in a particular
(assumed) causal structure. Simple maximum likelihood estimation of these strength
parameters does not capture the trends in the data, and so we have considered
estimation algorithms that do not have a strong rational justification. Weare thus
led to the question of whether human learning curves can be accounted for by a
rational process. In this section, we argue that the key to forming a rational,
statistical explanation of people's dynamical behavior is to take structural
uncertainty into account when forming parameter estimates.
Complete specification of the structure of a Bayesian network includes both the
underlying graph and choice of parameterization. For example, in the present task
there are three possible relationships between a potential cause C1 and an. effect E:
generative (h+), preventive (h_), or non-existent (h o). These three possibilities can
respectively be represented by a graph with a noisy-OR parameterization, one with a
noisy-AND-NOT parameterization, and one with no edge between the potential
cause and the effect. Each possibility is illustrated schematically in Figure 3.
ho
~@
~-~
?
?
+~N;_
Figure 3: Structural hypotheses for the Bayesian model. Co is an always present
background cause, C1 is the potential cause, and E the effect. The signs of arrows
indicate positive and negative influences on the outcome.
Previous work applying Bayesian structure learning to human causal judgment
focused on people .making the decision as to which of these structures best accounts
for the observed data [9]. That work showed that the likelihood of finding a causal
relationship rose with the base rate peE) in non-contingent cases, suggesting that
structural decisions are a relevant part of the present data. However, the rating scale
of the current task seems to encourage strength judgments rather, than purely
structural decisions, because it is anchored at the endpoints by two qualitatively
different causal strengths (strong generative, strong preventive). As a result,
subj ects' causal judgments appear to converge to causal power.
Real causal learning tasks often involve uncertainty about both structure and
parameters. Thus, even when a task demands ratings of causal strength, the
structural uncertainty should still be taken into account; we do this by considering a
hierarchy of causal models. The first level of this hierarchy involves structural
uncertainty, giving equal probability to the relationship between the variables being
generative, preventive, or non-existent. As mentioned in previous sections, the
parameterizations associated with the first two models lead to' a maximum
likelihood estimate of causal power. The second level of the hierarchy addresses
uncertainty over the parameters. With a constant background and a single cause,
there are two parameters for the noisy-OR and the noisy-AND-NOT models, Va and VI. If the cause and effect are unconnected, then only Va is required. Uncertainty in
all parameters can be expressed with distributions on the unit interval.
Using this set of m9dels, we can obtain a strength rating by taking the expectation
of the strength parameter Vi associated with a causal variable over the posterior
distribution on that parameter induced by the data. This expectation is taken over
both structure and parameters, allowing both factors to influence the result. In the
two-variable case, we can write this as
1
<11 >== L J11 "111 h,D) "hi D)dW
hEHO
where H = {h+, ha, h_}. The effective value of the strength parameter is a in the
model where there is no relationship between cause and effect, and should be
negative for preventive causes. We thus have:
<VI> = P(h+)f.l+ - P(h_)f.l-
where f.l+, f.l- are the posterior means of VI under h+ and h_ respectively.
While this theory is appealing from a rational and statistical point of view, it has
computational drawbacks. All four terms in the above expression are quite
computationally intensive to compute, and require an amount of information that
increases exponentially with the number of causes. Furthermore, the number of
different hypotheses we must consider grows exponentially with the number of
potential causes, limiting its applicability for multivariate cases.
We applied this model to the data of [7J, using a uniform prior over models, and
also over parameters. The results, averaged across 200 random orderings of trials,
are shown in Figure 2 (e). The predictions are somewhat symmetric with respect to
positive and negative contingencies and high and low peE). This symmetry is a
consequence of choosing a uniform (i.e., strongly uninformative) prior for the
parameters. If we instead take a uniform prior on VI and a beta(1,5) prior on Va,
consistent with a prior belief that effects occur only rarely without an observed
cause and similar to starting with zero weights in the algorithms presented above,
we obtain the results shown in Figure 2 (t). In both cases, the curvature of the
learning curves is a consequence of structural uncertainty, and the asymptotic values
reflect the strength of causal relationships. In the contingent cases, the probability
distribution over structures rapidly transfers all of its mass to the correct hypothesis,
and the result asymptotes at the posterior mean of' VI in that model, which will be
very close to causal power. The initial non-zero ratings in the non-contingent cases
result from h+ giving a slightly better account of the data than h_, essentially due to
the non-uniform prior on Va.
This structural account is only one means of understanding the rational basis for
these learning curves. Dayan and Kakade [3] provide a statistical theory of classical
conditioning based on Bayesian estimation of the parameters in a linear model
similar to that underlying 11P. Their theory accounts for phenomena that the
classical R-W theory does not, such as backwards blocking. They also give a neural
network learning model that approximates the Bayesian estimate, and that closely
resembles the augmented R-W model considered here. Their network model can
also produce the learning curves discussed in this paper. However, because it is
based on a linear model of causal interaction, it is not a good candidate for modeling
human causal judgments, which across various studies of asymptotic behavior seem
to be more closely approximated by parameter estimates' in noisy logic gates, as
instantiated in the power PC model [1] and our Bayesian model.
5
Conclusion
In this paper, we have outlined a range of dynamical models, from computationally
simple ones (such as simply applying conditional liP to the observed datapoints) to
rationally grounded ones (such as Bayesian structure/parameter estimation).
Moreover, there seems to be a tension in this domain in trying to develop a model
that is easily implemented in an individual and scales well with additional variables,
and one that has a rational statistical basis. Part of our effort here has been aimed at
providing a set of models that seem to equally well explain human behavior, but that
have different virtues besides their fit with the data. Human causal learning might
not scale up well, or it might not be rational; further discrimination among these
possible theories awaits additional data about causal learning curves.
References
[1] Cheng, Patricia W. 1997. "From Covariation to Causation: A Causal Power Theory."
Psychological Review, 104 (2): 367-405.
[2] Danks, David. Forthcoming. "Equilibria of the Rescorla-Wagner Model." Journal of
Mathematical Psychology.
[3] Dayan, Peter, & Kakade, Sham. 2001. "Explaining Away in Weight Space." In Advances
in Neural Information Processing Systems 13.
[4] Gopnik, Alison, Clark Glymour, David M. Sobel, Laura E. Schulz, Tamar Kushnir, &
David Danks. 2002. "A Theory of Causal Learning in Children: Causal Maps and Bayes
Nets." Submitted to Psychological Review.
[5] Lober, Klaus, & David R. Shanks. 2000. "Is Causal Induction Based on Causal Power?
Critique of Cheng (1997)." Psychological Review, 107 (1): 195:..212.
[6] Rescorla, Robert A., & Allan R. Wagner. 1972. "A Theory of Pavlovian Conditioning:
Variations in the Effectiveness of Reinforcement and Nonreinforcement." In A. H. Black &
W. F. Prokasy, eds. Classical Conditioning II: Current Research and Theory. New York:
Appleton-Century-Crofts. pp. 64-99.
[7] Shanks, David R. 1995. "Is Human Learning Rational?" The Quarterly Journal of
Experimental Psychology, 48A (2): 257-279.
[8] Spellman, Barbara A. 1996. "Conditionalizing Causality." In D. R. Shanks, K. J.
Holyoak, & D. L. Medin, eds. 1996., Causal Learning: The Psychology of Learning and
Motivation, Vol. 34. San Diego, Calif.: Academic Press. pp. 167-206.
[9] Tenenbaum, Joshua B., & Thomas L. Griffiths. 2000. "Structure Learning in Human
Causal Induction." In Advances in Neural Information Processing Systems 13.
[10] Van Hamme, Linda J., & Edward A. Wasserman. 1994. "Cue Competition in Causality
Judgments: The Role of Nonpresentation of Compound Stimulus Elements." Learning and
Motivation, 25: 127-151.
| 2158 |@word trial:6 version:2 proportion:1 seems:2 holyoak:1 gradual:2 simulation:1 simplifying:1 initial:7 score:2 past:1 current:3 comparing:1 must:1 readily:1 confirming:1 asymptote:7 discrimination:1 generative:12 cue:1 device:1 parameterization:5 short:2 characterization:1 parameterizations:2 five:1 mathematical:1 constructed:1 c2:7 differential:1 beta:2 ect:2 qualitative:4 f1p:4 combine:2 allan:1 indeed:1 behavior:11 nor:4 brain:1 little:1 actual:1 considering:1 increasing:1 estimating:1 underlying:3 matched:1 moreover:1 mass:1 linda:1 what:2 psych:1 ail:2 orland:1 unobserved:1 finding:1 quantitative:1 every:1 interactive:1 exactly:2 unit:1 appear:1 producing:1 arguably:1 positive:5 before:1 understood:1 limit:2 consequence:2 despite:1 critique:1 might:3 black:1 au:1 resembles:1 co:2 range:2 medin:1 averaged:1 disconfirming:1 implement:1 griffith:2 regular:2 wait:1 suggest:1 cannot:2 close:1 applying:3 influence:2 map:1 attention:1 regardless:2 starting:1 focused:1 simplicity:1 immediately:1 wasserman:1 estimator:7 rule:2 datapoints:2 dw:1 century:1 variation:3 justification:3 limiting:1 target:1 gm:1 suppose:1 hierarchy:3 diego:1 us:2 hypothesis:3 trend:1 fjo:1 approximated:1 element:1 updating:1 predicts:4 blocking:3 observed:9 role:1 capture:2 descend:1 ordering:1 decrease:1 principled:2 rose:1 mentioned:1 moderately:1 asked:1 dynamic:5 existent:2 raise:1 purely:1 upon:1 basis:3 po:3 easily:2 represented:1 various:1 instantiated:1 effective:1 klaus:1 choosing:2 outcome:1 quite:1 stanford:3 otherwise:1 jud:1 favor:1 statistic:2 noisy:8 final:1 sequence:1 advantage:1 net:5 rescorla:3 interaction:1 relevant:1 date:1 rapidly:1 conditionalizing:1 competition:1 ent:1 convergence:4 produce:7 converges:3 depending:1 develop:1 measured:1 received:1 edward:1 strong:4 implemented:1 predicted:1 involves:2 judge:1 indicate:1 convention:1 drawback:1 correct:2 closely:2 gopnik:1 camouflage:2 human:13 require:1 camouflaged:1 considered:3 normal:1 equilibrium:2 cognition:1 predict:1 early:1 pee:16 perceived:1 estimation:6 currently:1 vice:1 clearly:1 danks:3 always:5 rather:2 focus:6 notational:1 properly:1 likelihood:14 contrast:7 flaw:1 dayan:2 lj:1 typically:1 schulz:1 tank:2 classification:1 among:1 denoted:1 equal:2 having:2 report:1 stimulus:1 h_:5 serious:1 primarily:1 escape:1 causation:1 randomly:1 individual:2 possibility:2 patricia:1 pc:11 undefined:1 sobel:1 edge:1 encourage:1 divide:1 calif:1 causal:49 theoretical:1 psychological:6 modeling:2 asking:1 aio:2 applicability:1 uniform:5 examining:1 reflect:2 satisfied:1 slowly:1 alo:2 cognitive:1 laura:1 li:1 account:8 potential:19 suggesting:1 includes:1 vi:12 performed:1 view:1 closed:1 observing:2 start:1 bayes:5 sort:1 hamme:1 likewise:1 judgment:11 correspond:1 yield:1 bayesian:11 accurately:1 submitted:1 explain:2 andnot:1 reach:1 suffers:1 ed:2 acquisition:2 pp:2 associated:2 static:1 rational:12 dataset:1 begun:1 covariation:1 cj:1 actually:1 i1p:3 appears:1 higher:1 follow:1 tension:1 response:1 though:3 strongly:1 furthermore:2 implicit:1 until:1 ei:2 marker:1 lack:1 behaved:1 grows:1 effect:16 true:1 unbiased:2 former:1 hence:1 symmetric:1 jbt:1 illustrated:1 during:1 trying:1 complete:1 vo:1 fj:1 rnit:1 endpoint:1 exponentially:2 conditioning:3 discussed:1 approximates:1 refer:1 cambridge:1 versa:1 ai:1 appleton:1 outlined:1 specification:1 surface:1 base:1 curvature:1 multivariate:2 posterior:3 recent:1 showed:1 barbara:1 compound:1 joshua:2 seen:1 nonreinforcement:1 contingent:20 somewhat:1 additional:2 determine:5 converge:2 ii:2 multiple:2 sham:1 reduces:1 match:1 academic:1 calculation:1 believed:1 long:5 divided:1 equally:1 va:5 prediction:8 variant:1 fjj:3 converging:1 essentially:3 circumstance:1 expectation:2 sometimes:2 grounded:1 c1:5 background:7 schematically:1 uninformative:1 addressed:1 interval:1 unlike:1 ascent:3 subject:3 induced:1 j11:1 seem:4 effectiveness:1 structural:11 lober:1 presence:2 backwards:3 destroyed:2 affect:1 fit:5 psychology:4 gave:1 forthcoming:1 opposite:1 idea:1 tamar:1 intensive:1 absent:3 shift:1 whether:7 motivated:1 expression:1 effort:1 peter:1 york:1 cause:42 jj:4 generally:1 involve:1 aimed:1 amount:2 tenenbaum:2 ten:1 shifted:1 sign:2 delta:2 estimated:3 write:1 vol:1 key:1 salient:2 four:1 changing:1 neither:1 asymptotically:1 graph:2 year:1 run:8 uncertainty:7 decision:3 fl:1 hi:1 shank:3 followed:2 cheng:3 strength:20 occur:1 subj:2 constraint:1 speed:1 simulate:1 gruffydd:1 pavlovian:1 format:1 conjecture:2 glymour:1 department:2 across:2 slightly:1 appealing:1 kakade:2 making:1 gradually:1 alison:1 taken:2 computationally:2 equation:6 agree:1 know:1 fji:1 awaits:1 multiplied:2 quarterly:1 away:1 appropriate:3 preventive:15 alternative:2 ho:1 gate:3 florida:1 existence:1 thomas:2 include:1 giving:2 classical:3 question:2 quantity:1 occurs:1 strategy:2 rationally:2 gradient:3 simulated:1 me:2 argue:1 induction:2 assuming:2 besides:1 relationship:6 ratio:1 providing:1 setup:1 robert:1 potentially:2 negative:7 implementation:1 weare:1 kushnir:1 perform:1 allowing:1 situation:2 unconnected:1 community:1 rating:11 david:6 pair:2 required:1 connection:1 raising:1 learned:1 address:1 able:1 dynamical:8 explanation:1 belief:1 analogue:2 power:25 representing:1 spellman:1 numerous:1 al0:1 irrespective:1 prior:8 understanding:1 review:3 ddanks:1 asymptotic:6 expect:1 proportional:4 clark:1 contingency:2 sufficient:1 consistent:3 reclassification:1 pi:3 summary:1 accounted:1 surprisingly:1 salience:2 institute:1 explaining:1 taking:1 wagner:3 absolute:2 van:1 curve:20 world:3 computes:2 preventing:1 made:2 qualitatively:2 reinforcement:1 san:1 far:1 prokasy:1 logic:2 active:1 sequentially:1 assumed:2 iterative:2 anchored:1 lip:1 transfer:1 ca:1 symmetry:1 interact:1 necessarily:1 domain:1 main:1 linearly:1 arrow:1 motivation:4 ait:1 child:1 augmented:14 causality:2 west:1 fails:1 candidate:1 third:2 croft:1 removing:1 normative:1 virtue:1 stepwise:1 sequential:2 adding:1 ci:2 downward:1 demand:1 led:1 simply:4 explore:1 likely:1 forming:2 desire:1 ordered:1 expressed:1 applies:1 ma:1 conditional:15 viewed:1 towards:1 absence:3 change:3 directionality:1 typical:1 infinite:2 specifically:1 except:1 called:1 experimental:3 rarely:1 people:7 latter:1 phenomenon:2 |
1,272 | 2,159 | Rational Kernels
Corinna Cortes Patrick Haffner Mehryar Mohri
AT&T Labs ? Research
180 Park Avenue, Florham Park, NJ 07932, USA
corinna, haffner, mohri @research.att.com
Abstract
We introduce a general family of kernels based on weighted transducers or rational relations, rational kernels, that can be used for analysis of
variable-length sequences or more generally weighted automata, in applications such as computational biology or speech recognition. We show
that rational kernels can be computed efficiently using a general algorithm of composition of weighted transducers and a general single-source
shortest-distance algorithm. We also describe several general families of
positive definite symmetric rational kernels. These general kernels can
be combined with Support Vector Machines to form efficient and powerful techniques for spoken-dialog classification: highly complex kernels
become easy to design and implement and lead to substantial improvements in the classification accuracy. We also show that the string kernels
considered in applications to computational biology are all specific instances of rational kernels.
1 Introduction
In many applications such as speech recognition and computational biology, the objects
to study and classify are not just fixed-length vectors, but variable-length sequences, or
even large sets of alternative sequences and their probabilities. Consider for example the
problem that originally motivated the present work, that of classifying speech recognition
outputs in a large spoken-dialog application. For a given speech utterance, the output of a
large-vocabulary speech recognition system is a weighted automaton called a word lattice
compactly representing the possible sentences and their respective probabilities based on
the models used. Such lattices, while containing sometimes just a few thousand transitions,
may contain hundreds of millions of paths each labeled with a distinct sentence.
The application of discriminant classification algorithms to word lattices, or more generally
weighted automata, raises two issues: that of handling variable-length sequences, and that
of applying a classifier to a distribution of alternative sequences. We describe a general
technique that solves both of these problems.
Kernel methods are widely used in statistical learning techniques such as Support Vector
Machines (SVMs) [18] due to their computational efficiency in high-dimensional feature
spaces. This motivates the introduction and study of kernels for weighted automata. We
present a general family of kernels based on weighted transducers or rational relations,
rational kernels which apply to weighted automata. We show that rational kernels can be
computed efficiently using a general algorithm of composition of weighted transducers and
a general single-source shortest-distance algorithm.
We also briefly describe some specific rational kernels and their applications to spokendialog classification. These kernels are symmetric and positive definite and can thus be
combined with SVMs to form efficient and powerful classifiers. An important benefit of
S EMIRING
Boolean
Probability
Log
Tropical
S ET
('*),+-/.,021
.,02354
Table 1: Semiring examples. ! is defined by: "#$ %&
.
our approach is its generality and its simplicity: the same efficient algorithm can be used
to compute arbitrarily complex rational kernels. This makes highly complex kernels easy
to use and helps us achieve substantial improvements in classification accuracy.
2 Weighted automata and transducers
In this section, we present the algebraic definitions and notation necessary to introduce
rational kernels.
-/6
4
-96
4
Definition 1 ([7]) A system
-9677 84 is a semiring if:
7 is a commutative
monoid with identity element ; 7 is a monoid
with
identity
element ; distributes
6
over ; and is an annihilator for : for all :<;
: !& =>:#& .
Thus, a semiring is a ring that may lack negation. Table 2 lists some familiar examples
of semirings. In addition to the Boolean semiring and the probability semiring used to
combine probabilities, two semirings often used in
' )+ applications are the log semiring which
is isomorphic to the probability semiring via a
morphism, and the tropical semiring
which is derived from the log semiring using the Viterbi approximation.
6
Definition
2 A weighted
transducer ? over a semiring is an 8-tuple ?@&
-BA
4 finite-state
A
CDEFGHJIKLMJN where: is the finite input alphabet of the transducer; C is the
FPXWOQ4 E the
finite output alphabet; E is a finite
set
-VA
XW of
4 states;
6 set of initial states; GROSE the
set of final6 states; I@OTEU
CY
E a finite set of transitions;
6
the initial weight function; and N]ZGU[
the final weight function mapping
LZF\6 [
G to .
Weighted automata can be formally defined in a similar way by simply omitting the input
or the output labels.
.
.cb
.cb
Given a transition ;^I , we denote
.cb by _a` its origin or previous
.hgjiciiJ.lstate
k and d` its destination state or next state, and e!.,` n g its
is an element of Im
b weight.
.5npb A path fQ&
d` b 0
&o._agJb` , q$&srtcuucuvw . We extend d and _ to paths
with consecutive
b transitions:
.kXb
by setting: d` f &xd`
and _a` f &o_a` . The weight function e can also be extended
to paths by defining
of its constituentz
b the. weight
i a path
. k bas the -product of-/the
z z|{}weights
4
g b iciof
!
e
`
f
s
&
!
e
`
Q
!
e
`
y
transitions:
.
We
denote
by
the
set
ofA paths from
z {
-9z
z {4
z
z {
to and by y
the set of paths from to with input label ";
~"a~%2
m { and output
label
case). These
- % (transducer
-/z definitions
z { 4 can be extended to subsets 7 O?E , by:
{4
y #J"8~%2
&?????j?V?~??????/y
~"a~%2 .
A transducer
is
regulated
if
the
output
weight associated by ? to any pair of input-output
?
4
string "a~% by:
`?
b V-
4
"8J%
6
&
?
b ?- 4
? ???? ?? 1 ? 3 ? ?M?
?b 4
b
bb
L a
_ `f ?
e!` f ^N2` d` f
-
(1)
4
is well-defined and in . ` ?
" & when y FJ"8~%2G &x? . In the following, we will
assume that all the transducers considered are regulated. Weighted transducers are closed
under , and
g Kleene-closure. In particular, - the 4 -sum and -multiplications of two
transducers ? and ?M? are defined for each pair "a~% by:
g
b ?4
gb V4
b V4
(2)
` ? ???? "8J%
&
`?
"8J% S` ` ?2? "8~%
`?
g
????
b ?-
"8J%
4
&
?
`?
1X?M1l?V1X? ? v3 ?M3v??3?
gb ?- g vg 4
b ?4
" J% ?
` ??? ?
" ?J%?
(3)
3 Rational kernels
This section introduces rational kernels, presents a general algorithm for computing them
efficiently and describes several examples of rational kernels.
3.1 Definition
Definition
3 A kernel4
-BA
CDEFA GHJIKLMJN
m:
all "a~%\;
is rational if 6 there exist a weighted
&
6 transducer ?
Z
[ such that for
over the semiring
and a function
-
4
"8J%
&
-
6
`?
b ?-
4~4
"a~%
(4)
In general, is an arbitrary function mapping to . In some cases, it may be desirable
to assume
that it is a semiring morphism as in Section 3.6.6 It is often the identity function
6
when &o and 6may be a projection
when the semiring is the cross-product of and
6 {
another semiring ( &S ?
).
Rational kernels can be naturally extended to kernels over weighted automata. In the
following, to simplify the presentation, we will restrict ourselves to the case of acyclic
weighted automata which is the case of interest for our applications, but our results apply
similarly to arbitrary
Let and be two acyclic weighted automata
6 weighted
- automata.
4
over the semiring , then
is defined by:
4 &
!
?
1 ?3
`
b V- 4
b V4
b V- 4J4
" S` ? 8
" ~% ?`
%
(5)
More generally, the results mentioned in the following for strings apply all similarly
to
6
acyclic weighted automata. Since the set of weighted transducers over a semiring is also
closed
under -sum and -product [2, 3], it follows that
6
g rational kernels over ag semiring
are closed under sum and product.
We
denote
by
? the sum and by
? the
g
g
product of two rational kernels
and <? . Let ? and ?M? be the associated transducers of
these kernels, we have for example:
-
g
?
4-
"a~%
4
&
- - g
V4 b ?~4 4
` ? ?? ? "a~% &
g -
"8J%
4
4
" ~%
? a
(6)
In learning techniques such as those based on SVMs, we are particularly interested in
positive definite symmetric kernels, which guarantee the existence of a corresponding reproducing kernel Hilbert space. Not all rational kernels are positive definite symmetric but
in the following sections we will describe some general classes of rational kernels that have
this property.
Positive definite symmetric kernels can be used to construct other families of kernels
that also meet these
conditions
[17]. Polynomial kernels of degree _ are
from
4
-~ formed
? ?? 4 with
the- expression
,
and
Gaussian
kernels
can
be
formed
as
T
:
4
4
4
? "a~% 4 &
"8J"
%2~%
r
"8J% . Since the class of symmetric positive definite kernels is closed under sum [1], the sum of two positive definite rational kernels is also
a positive definite rational kernel.
In what follows, we will focus
for computing rational kernels. The al4 on the - algorithm
4
"a~% , or ! , for any two acyclic weighted automata, is
gorithm for computing
based on two general algorithms that we briefly present: composition of weighted trans6
ducers to combine , ? , and , and a general shortest-distance algorithm in a semiring
to compute the -sum of the weights of the successful paths of the combined machine.
3.2 Composition of weighted transducers
Composition is a fundamental operation on weighted transducers that can be used
in many
6
applications to create complex
g weighted transducers from simpler ones. Let 6 be a commutative semiring and let ? and ?M? be two weighted transducersg defined over such that
the input
alphabet of ? ? coincides withg the output alphabet of ? . Then, the composition
g
of ? and ?M? is a weighted transducer ?
??? which, when it is regulated, is defined for all
a:a/1.61
0
1
a:b/0
a:a/1.2
b:a/0.69
b:a/0.69
2
b:b/0.22
0
3/0
b:b/0.92
a:b/2.3
b:a/0.51
(a)
0
a:a/0.51
1
2/0
(b)
a:a/2.81
1
a:a/0.51
a:b/3.91
4
a:b/0.92
b:a/1.2
2
3/0
b:a/0.73
(c)
g
Figure 1: (a) Weighted transducer ? over the log semiring. (b) Weighted
transducer ? ?
g
over the log semiring. (c) Construction of the result of composition ?
? ? . Initial states
are represented by bold circles, final states by double circles. Inside each circle, the first
number indicates the state number, the second, at final states only, the value of the final
weight function N at that state. Arrows represent transitions and are labeled with symbols
followed by their corresponding weight.
"8J% by [2, 3, 15, 7]: 1
`?
g
???
b V-
"8J%
4
&??
gb V-
`?
"8
4
?` ???
b ?-
4
~%
(7)
A
Note that a transducer can be viewed as a matrix over a countable set m!
CKm and composition as the corresponding matrix-multiplication. There exists a general and efficient
composition algorithm for weighted transducers which takesg advantage of the sparsity of
the inputg transducers [14, 12]. States in the compositiong ?
??? of two weighted transducers ? and ?8W ? are identified with pairs of a state of ? and a state of ?8? . Leaving aside
transitions
with inputs or outputs, the following
rule specifies how to compute a transition
g
g
of ?
? ? from appropriate transitions of ? and ? ? :2
-9z g
J:? ~e
g z 4
?
-/z g{
z {4
-~-9z g z g{ 4
g
-/z
z { 4J4
and
(8)
~e ? ? &
:hl~e ?e ? ? ?
g
z g
In
zlg{ the worst case, all transitions of ? leaving a state match all those
-~-
ofg ?a
?
leaving
g
4-
state
E
I
E ?
, thus
4~4 the space and time complexity of composition is quadratic:
I ? . Fig.(1) (a)-(c) illustrate the algorithm when applied to the transducers of Fig.(1) (a)(b) defined over the log semiring. The intersection of two weighted automata is a special
case of composition. It corresponds to the case where the input and output label of each
transition are identical.
3.3 Single-source shortest distance algorithm over a semiring
Given a weighted automaton or transducer , the shortest-distance
from state
z
of final states G is defined as the -sum of all the paths from to G :
` zXb &
?
e `f
!
? ? ???}?? ?M?
6
b
?N2` d` f
b*b
z
to the set
(9)
when this sum is well-defined and in , which is always the case when the semiring is w closed or when is acyclic [11], the case of interest
zXb in what follows.-
There
- exists a general
4
4
algorithm for computing the shortest-distance ` in linear time E ?
(?? I ,
where ? denotes the maximum time to compute and ? the time to compute [11].
The algorithm
is a generalization of Lawler?s algorithm [8] to the case of an arbitrary
6
semiring . It is based on a generalized relaxation of the outgoing transitions of each
state of visited in reverse topological order [11].
1
We use a matrix notation for the definition of composition as opposed to a functional notation.
This is a deliberate choice motivated by an improved readability in many applications.
2
See [14, 12] for a detailed presentation of the algorithm including the use of a transducer filter
for dealing with -multiplicity in the case of non-idempotent semirings.
?:b/3
?:a/3
b:?/2
a:?/2
b:a/1
a:b/1
b:b/0
a:a/0
?:b
?:a
b:?
a:?
?:a
a:a b:?/? ?:a/? 3
a:?/?
?:b/?
a:a
2
b:b
1
?:b
b:b
a:a
b:b
0
0/0
?:b/?
?:a/?
(a)
a:a
b:b
?:b
?:a
b:?
a:?
4
?:a
?:b
5
(b)
Figure 2: Weighted transducers associated to two rational kernels. (a) Edit-distance kernel.
(b) Gappy -gram count kernel, with = 2.
3.4 Algorithm
Let
be a rational kernel and let ? be the associated weighted transducer. Let and
A
be two acyclic weighted automata.
and may represent just two strings "8J%^;
m or
may be any other complex weighted - acceptors.
By definition of rational kernels (Eq.(5))
4
can be computed by:
and the shortest-distance (Eq.(9)),
!
.
1. Constructing theb acyclic composed transducer &
?
2. Computing ` , the shortest-distance from the initial states of to its final states
using the shortest-distance algorithm described in the previous section.
b?4
.
3. Computing
`
-
4
Thus, the total complexity of the algorithm is ?
, where ? ,
, and
denote
the size of ? , and and the worst case complexity of computing
- 4 respectively
6
time as in many applica" , "?; . If we assume that can be computed- in constant
4
tions, then the
of the computation of
is quadratic with respect to
-
complexity
4
and is: ?
.
3.5 Edit-distance kernels
Recently, several kernels, string kernels, have been introduced in computational biology for
input vectors representing biological sequences [4, 19]. String kernels are specific instances
of rational kernels. Fig.(2) (a) shows the weighted transducer over the tropical semiring
associated to a classical type of string kernel. The kernel corresponds to an edit-distance
based on a symbol substitution with cost , deletion with cost r , and insertion of cost
. All classical edit-distances can be represented by weighted transducers over the tropical
semiring [13, 10]. The kernel computation algorithm just described can be used to compute
efficiently the edit-distance of two strings or two sets of strings represented by automata. 3
3.6 Rational kernels of the type ?
?
0 g
There exists a general method for constructing a 6positive definite and symmetric rational
kernel from a weighted transducer
? when Z
[ is
6
0 g a semiring morphism ? this
implies in particular that is commutative. Denote by ?
the inverse of ? , that is the
g
transducer obtained from ? by transposing
the
input
and
output
labels of each transition.
0
Then the composed transducer &S? ?
is symmetric and, when it is regulated, defines
3
We have proved and will present elsewhere a series of results related to kernels based on the
notion of edit-distance. In particular, we have shown that the classical edit-distance with equal
costs for insertion, deletion and substitution is not negative definite [1] and that the Gaussian kernel
is not positive definite.
a positive definite symmetric rational kernel
by definition of composition:
-
"8J%
- `
4
&
b ?-
"a~%
4~4
. Indeed, since
&
-
`?
b ?-
"8
4J4?i
-
is a semiring morphism,
`?
b ?-
%2
4J4
which shows that is symmetric. For any non-negative integer d and for all "a~% we define
a symmetric kernel
by:
- "8~% 4 & - ` ? b ?- "8 4J4?i - ` ? b V- %? 4J4
g
? ucuuv A
where the sum runs over all strings of length less or equal to d . Let
g
be an arbitrary ordering of nthese
For any
and any " uucuJ"
;
m,
& strings.
define
by:
- " n ~" X4 . Then, & with defined by n &
- b ?- then matrix
X 4J4
` ? " . Thus, the eigenvalues of are all non-negative,
- which
4
' implies that - is
4
"8J% &
"8J% ,
positive definite [1]. Since is a point-wise limit of ,
is also definite positive [1].
4 Application to spoken-dialog classification
Rational kernels can be used in a variety of applications ranging from computational biology to optical character recognition. This section singles out one specific application, that
of topic classification applied to the output of a speech recognizer. We will show how the
use of weighted transducers rationalizes the design and optimization of kernels. Simple
equations and graphs replace complex diagrams and intricate algorithms often used for the
definition and analysis of string kernels.
As mentioned in the introduction, the output of a speech recognition system associated
to a speech utterance is a weighted automaton called a word lattice representing a set of
alternative sentences and their respective probabilities based on the models used. Rational
kernels help address both the problem of handling variable-length sentences and that of
applying a classification algorithm to such distributions of alternatives.
The traditional solution to sentence classification is the ?bag-of-words? approach used in
information retrieval. Because of the very large dimension of the input space, the use of
large-margin classifiers such as SVMs [6] and AdaBoost [16] was found to be appropriate
in such applications.
One approach adopted in various recent studies to measure the topic-similarity of two sentences consists of counting their common non-contiguous -grams, i.e., their common
substrings of
words with possible insertions. These -grams can be extracted explicitly from each sentence [16] or matched implicitly through a string kernel [9]. We will
show that such kernels are rational and will describe how they can be easily constructed
and computed using the general algorithms given in the previous section. More generally,
we will show how rational kernels can be used to compute the expected counts of common
non-contiguous -grams of two weighted automata and thus define the topic-similarity of
two lattices. This will demonstrate the simplicity, power, and flexibility of our framework
for the design of kernels.
4.1 Application of ?
?
0 g
kernels
Consider a word lattice over the probability
semiring. can be viewed as a probability
A
distribution y
over all strings <;
m . The expected count or number of occurrences
- 4
of
1 ,
"
y
?
y
an -gram
sequence
in
a
string
for
the
probability
distribution
is:
where 1 denotes the number of occurrences of " in . It is easy to construct a weighted
transducer ?
that outputs the set of -grams of an input lattice with their corresponding
counts. Fig.(3) (a) shows that transducer, when the alphabet is reduced to
A expected
& :? and &?r . Similarly, the transducer ? H? of Fig.(3) (b) can be used to output
non-contiguous or gappy -grams with their expected counts. 4 Long gaps are penalized
!#" $
%
4
The transducers shown in the figures of this section are all defined over the probability semiring,
thus a transition corresponding to a gap in
is weighted by .
b:?
a:?
a:a
b:b
0
b:?
a:?
b:?
a:?
2
0
a:a
b:b
1
b:?
a:?
b:?/?
a:?/?
a:a
b:b
1
(a)
a:a
b:b
2
(b)
Figure 3: -gram transducers ( = 2) defined over the probability semiring. (a) Bigram
counter transducer ?a? . (b) Gappy bigram counter ??v? .
with a decay factor ?L Y : a gap of length reduces the count by L . A transducer
counting variable-length -grams is obtained by simply taking the sum of these transducers: ?
? &
? t? .
and L since our results are
In the remaining of this section, we will omit the subscript
independent of the choice of these parameters. Thus the topic-similarity of two strings or
lattices and based on the expected counts of theirs g common substrings is given by:
-
4
&
`
-
?
?
0 4
b
(10)
is of the type studied in section 3.6 and thus is symmetric and positive
The kernel
definite.
4.2 Computation
The specific form of the kernel
and the associativity of composition provide us with
several alternatives for computing .
General algorithm. We can use the general
0 g algorithm described in Section 3.4 to compute
by precomputing the transducer ? ?
. Fig.(2)(b) shows the result of that composition
in the case of - gappy4 bigrams. Using that algorithm, the complexity of the -computation
of
4
the kernel
as described in the previous section is quadratic
. This
particular example has been treated by ad hoc algorithms with a similar complexity, but
that only work with strings [9, 5] and not with weighted automata or lattices.
Other factoring. Thanks to the associativity of composition, we can consider a different
factoring of the composition cascade defining :
g
-
4
&
`-
?
0 g
4
- 0
?
4Vb
first and then composing the resulting
This factoring suggests computing ? and0 ?g
transducers rather than constructing ?
. The choice between the two methods does
?
(11)
not affect the overall time complexity of the algorithm, but in practice one method may be
preferable over the other. We are showing elsewhere that in the specific case of the counting
transducers such as those described in previous
the kernel computation can in fact
-
sections,
4
be performed in linear time, that is in
, in particular by using the notion of
failure functions.
4.3 Experimental results
0 g
We used the ?
-type kernel with SVMs for call-classification in the spoken language
?
understanding (SLU) component of the AT&T How May I Help You natural dialog system.
In this system, users ask questions about their bill or calling plans and the objective is to
assign a class to each question out of a finite set of 38 classes made of call-types and named
entities such as Billing Services, or Calling Plans.
In our experiments, we used 7,449 utterances as our training data and 2,228 utterances as
our test data. The feature space corresponding to our lattice kernel is that of all possible
trigrams over a vocabulary of 5,405 words. Training required just a few minutes on a single
processor of a 1GHz Intel Pentium processor Linux cluster with 2GB of memory and 256
KB cache. The implementation took only about a few hours and was entirely based on
the FSM library. Compared to the standard approach of using trigram counts over the
best recognized sentence, our experiments with a trigram rational kernel showed a
reduction in error rate at a rejection level.
5 Conclusion
In our classification experiments in spoken-dialog applications, we found rational kernels
to be a very powerful exploration tool for constructing and generalizing highly efficient
string and weighted automata kernels. In the design of learning machines such as SVMs,
rational kernels give us access to the existing set of efficient and general weighted automata
algorithms [13]. Prior knowledge about the task can be crafted into the kernel using graph
editing tools or weighted regular expressions, in a way that is often more intuitive and easy
to modify than complex matrices or formal algorithms.
References
[1] Christian Berg, Jens Peter Reus Christensen, and Paul Ressel. Harmonic Analysis on Semigroups. Springer-Verlag: Berlin-New York, 1984.
[2] Jean Berstel. Transductions and Context-Free Languages. Teubner Studienbucher: Stuttgart,
1979.
[3] Samuel Eilenberg. Automata, Languages and Machines, volume A-B. Academic Press, 1974.
[4] David Haussler. Convolution kernels on discrete structures. Technical Report UCSC-CRL-9910, University of California at Santa Cruz, 1999.
[5] Ralf Herbrich. Learning Kernel Classifiers. MIT Press, Cambridge, 2002.
[6] Thorsten Joachims. Text categorization with support vector machines: learning with many
relevant features. In Proc. of ECML-98. Springer Verlag, 1998.
[7] Werner Kuich and Arto Salomaa. Semirings, Automata, Languages. Number 5 in EATCS
Monographs on Theoretical Computer Science. Springer-Verlag, Berlin, Germany, 1986.
[8] Eugene L. Lawler. Combinatorial Optimization: Networks and Matroids. Holt, Rinehart, and
Winston, 1976.
[9] Huma Lodhi, John Shawe-Taylor, Nello Cristianini, and Christopher J. C. H. Watkins. Text
classification using string kernels. In NIPS, pages 563?569, 2000.
[10] Mehryar Mohri. Edit-Distance of Weighted Automata. In Jean-Marc Champarnaud and Denis
Maurel, editor, Seventh International Conference, CIAA 2002, volume to appear of Lecture
Notes in Computer Science, Tours, France, July 2002. Springer-Verlag, Berlin-NY.
[11] Mehryar Mohri. Semiring Frameworks and Algorithms for Shortest-Distance Problems. Journal of Automata, Languages and Combinatorics, 7(3):321?350, 2002.
[12] Mehryar Mohri, Fernando C. N. Pereira, and Michael Riley. Weighted automata in text and
speech processing. In ECAI-96 Workshop, Budapest, Hungary. ECAI, 1996.
[13] Mehryar Mohri, Fernando C. N. Pereira, and Michael Riley. The Design Principles of a
Weighted Finite-State Transducer Library. Theoretical Computer Science, 231:17?32, January
2000. http://www.research.att.com/sw/tools/fsm.
[14] Fernando C. N. Pereira and Michael D. Riley. Speech recognition by composition of weighted
finite automata. In Emmanuel Roche and Yves Schabes, editors, Finite-State Language Processing, pages 431?453. MIT Press, Cambridge, Massachusetts, 1997.
[15] Arto Salomaa and Matti Soittola. Automata-Theoretic Aspects of Formal Power Series.
Springer-Verlag: New York, 1978.
[16] Robert E. Schapire and Yoram Singer. Boostexter: A boosting-based system for text categorization. Machine Learning, 39(2/3):135?168, 2000.
[17] Bernhard Scholkopf and Alex Smola. Learning with Kernels. MIT Press: Cambridge, MA,
2002.
[18] Vladimir N. Vapnik. Statistical Learning Theory. John Wiley & Sons, New-York, 1998.
[19] Chris Watkins. Dynamic alignment kernels. Technical Report CSD-TR-98-11, Royal Holloway,
University of London, 1999.
| 2159 |@word cu:1 briefly:2 bigram:3 polynomial:1 lodhi:1 closure:1 tr:1 reduction:1 initial:4 substitution:2 series:2 att:2 existing:1 com:2 john:2 cruz:1 applica:1 christian:1 aside:1 boosting:1 readability:1 herbrich:1 denis:1 simpler:1 constructed:1 ucsc:1 become:1 scholkopf:1 transducer:49 consists:1 combine:2 inside:1 introduce:2 intricate:1 expected:5 indeed:1 dialog:5 cache:1 notation:3 matched:1 what:2 string:19 kleene:1 jik:2 spoken:5 ag:1 monoid:2 nj:1 guarantee:1 ofa:1 xd:1 preferable:1 classifier:4 omit:1 appear:1 positive:14 service:1 modify:1 limit:1 meet:1 path:9 subscript:1 studied:1 suggests:1 practice:1 definite:15 implement:1 cascade:1 acceptor:1 projection:1 word:7 regular:1 holt:1 context:1 applying:2 www:1 bill:1 automaton:29 simplicity:2 rinehart:1 rule:1 haussler:1 ralf:1 notion:2 construction:1 user:1 origin:1 element:3 recognition:7 particularly:1 gorithm:1 labeled:2 ckm:1 worst:2 thousand:1 cy:1 ordering:1 counter:2 substantial:2 mentioned:2 monograph:1 complexity:7 insertion:3 cristianini:1 dynamic:1 raise:1 efficiency:1 gu:1 compactly:1 easily:1 represented:3 various:1 alphabet:5 distinct:1 describe:5 london:1 jean:2 widely:1 florham:1 schabes:1 final:6 hoc:1 sequence:7 advantage:1 eigenvalue:1 took:1 product:5 relevant:1 budapest:1 hungary:1 flexibility:1 achieve:1 boostexter:1 intuitive:1 double:1 cluster:1 categorization:2 ring:1 object:1 help:3 illustrate:1 tions:1 ij:1 eq:2 solves:1 implies:2 filter:1 kb:1 exploration:1 assign:1 generalization:1 biological:1 considered:2 cb:3 viterbi:1 mapping:2 lm:2 trigram:3 consecutive:1 recognizer:1 proc:1 bag:1 label:5 combinatorial:1 visited:1 edit:8 create:1 tool:3 weighted:53 mit:3 gaussian:2 always:1 rather:1 derived:1 focus:1 joachim:1 improvement:2 gappy:3 fq:1 indicates:1 pentium:1 factoring:3 associativity:2 eatcs:1 relation:2 france:1 interested:1 germany:1 issue:1 classification:12 overall:1 plan:2 special:1 equal:2 construct:2 biology:5 identical:1 park:2 x4:1 report:2 simplify:1 few:3 npb:1 composed:2 familiar:1 semigroups:1 ourselves:1 transposing:1 negation:1 interest:2 highly:3 alignment:1 introduces:1 semirings:4 berstel:1 tuple:1 fsm:2 necessary:1 respective:2 taylor:1 circle:3 theoretical:2 instance:2 classify:1 rationalizes:1 boolean:2 contiguous:3 werner:1 lattice:10 riley:3 cost:4 subset:1 tour:1 hundred:1 successful:1 seventh:1 combined:3 thanks:1 fundamental:1 international:1 destination:1 v4:4 michael:3 roche:1 linux:1 containing:1 opposed:1 bold:1 combinatorics:1 explicitly:1 ad:1 performed:1 teubner:1 lab:1 closed:5 formed:2 yves:1 accuracy:2 efficiently:4 kxb:1 substring:2 processor:2 definition:10 failure:1 naturally:1 associated:6 rational:38 proved:1 massachusetts:1 ask:1 knowledge:1 hilbert:1 lawler:2 originally:1 adaboost:1 improved:1 editing:1 stuttgart:1 generality:1 just:5 smola:1 tropical:4 christopher:1 lack:1 defines:1 ucu:2 usa:1 omitting:1 contain:1 ofg:1 symmetric:12 coincides:1 samuel:1 generalized:1 m1l:1 theoretic:1 demonstrate:1 gh:2 ranging:1 wise:1 harmonic:1 recently:1 common:4 functional:1 volume:2 million:1 extend:1 theirs:1 composition:18 cambridge:3 uv:1 similarly:3 language:6 shawe:1 access:1 j4:7 similarity:3 patrick:1 showed:1 recent:1 reverse:1 verlag:5 arbitrarily:1 jens:1 recognized:1 shortest:10 v3:1 fernando:3 july:1 desirable:1 reduces:1 arto:2 technical:2 match:1 academic:1 cross:1 long:1 retrieval:1 va:1 kernel:83 sometimes:1 represent:2 addition:1 diagram:1 source:3 leaving:3 integer:1 call:2 counting:3 easy:4 variety:1 affect:1 restrict:1 identified:1 billing:1 haffner:2 avenue:1 motivated:2 lstate:1 expression:2 and0:1 gb:1 peter:1 algebraic:1 speech:10 york:3 generally:4 detailed:1 santa:1 svms:6 reduced:1 http:1 specifies:1 schapire:1 exist:1 deliberate:1 discrete:1 graph:2 relaxation:1 sum:11 run:1 inverse:1 powerful:3 you:1 named:1 family:4 vb:1 entirely:1 followed:1 quadratic:3 topological:1 winston:1 alex:1 calling:2 aspect:1 optical:1 describes:1 son:1 character:1 christensen:1 multiplicity:1 thorsten:1 equation:1 count:8 singer:1 adopted:1 operation:1 apply:3 appropriate:2 occurrence:2 alternative:5 corinna:2 jn:2 existence:1 denotes:2 remaining:1 sw:1 xw:1 yoram:1 emmanuel:1 classical:3 objective:1 question:2 traditional:1 regulated:4 distance:18 berlin:3 entity:1 chris:1 topic:4 ressel:1 nello:1 discriminant:1 length:8 vladimir:1 robert:1 negative:3 ba:3 design:5 countable:1 motivates:1 implementation:1 convolution:1 finite:9 ecml:1 january:1 defining:2 extended:3 reproducing:1 arbitrary:4 introduced:1 david:1 semiring:33 pair:3 required:1 sentence:8 california:1 deletion:2 huma:1 hour:1 nip:1 address:1 sparsity:1 including:1 memory:1 royal:1 power:2 treated:1 natural:1 representing:3 library:2 idempotent:1 utterance:4 slu:1 text:4 prior:1 understanding:1 eugene:1 multiplication:2 lecture:1 acyclic:7 vg:1 degree:1 principle:1 editor:2 classifying:1 cd:2 elsewhere:2 mohri:6 penalized:1 free:1 ecai:2 formal:2 taking:1 matroids:1 benefit:1 ghz:1 dimension:1 vocabulary:2 transition:14 gram:9 made:1 ici:1 bb:1 implicitly:1 bernhard:1 dealing:1 table:2 matti:1 composing:1 mehryar:5 complex:7 constructing:4 marc:1 precomputing:1 csd:1 arrow:1 paul:1 n2:2 fig:6 intel:1 crafted:1 transduction:1 ny:1 wiley:1 morphism:4 pereira:3 watkins:2 minute:1 specific:6 showing:1 symbol:2 list:1 decay:1 cortes:1 exists:3 workshop:1 vapnik:1 commutative:3 margin:1 gap:3 rejection:1 intersection:1 generalizing:1 simply:2 salomaa:2 springer:5 corresponds:2 extracted:1 ma:1 identity:3 presentation:2 viewed:2 srt:1 replace:1 crl:1 distributes:1 called:2 total:1 isomorphic:1 experimental:1 formally:1 holloway:1 berg:1 support:3 outgoing:1 handling:2 |
1,273 | 216 | Rule Representations in a Connectionist Chunker
Rule Representations in a Connectionist Chunker
David S. Touretzky
Gillette Elvgren
School of Computer Science
Carnegie Mellon University
Pittsburgh, PA 15213
m
ABSTRACT
We present two connectionist architectures for chunking of symbolic
rewrite rules. One uses backpropagation learning, the other competitive
learning. Although they were developed for chunking the same sorts
of rules, the two differ in their representational abilities and learning
behaviors.
1 INTRODUCTION
Chunking is a process for generating, from a sequence of if-then rules, a more complex
rule that accomplishes the same task in a single step. It has been used to explain incremental human perfonnance improvement in a wide variety of cognitive, perceptual, and
motor tasks (Newell, 1987). The SOAR production system (Laird, Newell, & Rosenbloom, 1987) is a classical AI computer program that implements a "unified theory of
cognition" based on chunking.
SOAR's version of chunking is a symbolic process that examines the working memory
trace of rules contributing to the chunk. In this paper we present two connectionist
rule-following architectures that generate chunks a different way: they use incremental
learning procedures to infer the environment in which the chunk should fire. The first
connectionist architecture uses backpropagation learning, and has been described previously in (Touretzky, 1989a). The second architecture uses competitive learning. It
exhibits more robust behavior than the previous one, at the cost of some limitations on
the types of rules it can learn.
The knowledge to be chunked consists of context-sensitive rewrite rules on strings. For
example, given the two rules
431
432
Touretzky and Elvgren
RI:
R2:
"change D to B when followed by E"
"change A to C when followed by B"
the model would go through the following derivation: ADE - (Rule RI) ABE - (Rule
R2) CBE. Rule RI's firing is what enables rule R2 to fire. The model detects this and
formulates a chunked rule (RI-R2) that can accomplish the same task in a single step:
R I-R2:
AD - CB I _ E
Once this chunk becomes active, the derivation will be handled in a single step, this way:
ADE - (Chunk RI-R2) CBE. The chunk can also contribute to the formation of larger
chunks.
2
CHUNKING VIA BACKPROPAGATION
Our first experiment, a three-layer backpropagation chunker, is shown in Figure 1. The
input layer is a string buffer into which symbols are shifted one at a time, from the right.
The output layer is a "change buffer" that describes changes to be made to the string.
The changes supported are deletion of a segment, mutation of a segment, and insertion
of a new segment. Combinations of these changes are also permitted.
Rules are implemented by hidden layer units that read the input buffer and write changes
(via their a connections) into the change buffer. Then separate circuitry, not shown in
the figure, applies the specified changes to the input string to update the state of the input
buffer. The details of this string manipulation circuitry are given in (Touretzky, 1989b;
Touretzky & Wheeler, 1990).
We will now go through the ADE derivation in detail. The model starts with an empty
input buffer and two rules: R I and R2.1 After shifting the symbol A into the input buffer,
no rule fires-the change buffer is all zeros. After shifting in the D, the input buffer
contains AD, and again no rule fires. After shifting in the E the input buffer contains
ADE, and rule R I fires, writing a request in the change buffer to mutate input segment 2
(counting from the right edge of the buffer) to a B. The input buffer and change buffer
states are saved in temporary buffers, and the string manipulation circuitry derives a new
input buffer state, ABE. This now causes rule R2 to fire. 2 It writes a request into the
change buffer to mutate segment 3 to a C. Since it was RI's firing that triggered R2,
the conditions exist for chunk formation. The model combines RI's requested change
with that of R2, placing the result in the "chunked change buffer" shown on the right in
Figure I. Backpropagation is used to teach the hidden layer that when it sees the input
buffer pattern that triggered RI (ADE in this case) it should produce via its f3 connections
the combined change pattern shown in the chunked change buffer.
The model's training is "self-supervised:" its own behavior (its history of rule firings)
is the source of the chunks it acquires. It is therefore important that the chunking
1 The initial rule set is installed by an external teacher using backpropagation.
2Note that Rl applies to positions 1 and 2 of the buffer (counting from the right edge), while R2 applies to
positions 2 and 3. Rules are represented in a position-independent manner, allowing them to apply anywhere
in the buffer that their environment is satisfied. The mechanism for achieving this is explained in (Touretzky.
1989a).
Rule Representations in a Connectionist Chunker
Chunked Change:
Change Buffer:
cur: [change seg. 3 to "C"]
[ change seg. 2 to "B" and
change seg . 3 to "C" ]
prey: [change seg . 2 to "B"]
next:
cur:
prey:
tc
B
E
IA IB I E I
A D E
Figure 1: Architecture of the backpropagation chunker.
process not introduce any behavioral errors during the intennediate stages of learning,
since no external teacher is present to force the model back on track should its rule
representations become corrupted. The original rules are represented in the a connections
and the chunked rules are trained using the j3 connections, but the two rule sets share the
same hidden units and input connections, so interference can indeed occur. The model
must actively preserve its a rules by continuous rehearsal: after each input presentation,
backpropagation learning on a contrast-enhanced version of the a change pattern is used
to counteract any interference caused by training on the j3 patterns. Eventually, when the
j3 weights have been learned correctly, they can replace the a weights.
The parameters of the model were adjusted so that the initial rules had a distributed
representation in the hidden layer, Le., several units were responsible for implementing
each rule. Analysis of the hidden layer representations after chunking revealed that the
model had split off some of the RI units to represent the RI-R2 chunk; the remainder
were used to maintain the original RI rule.
The primary flaw of this model is fragility. Constant rehearsal of the original rule set, and
low learning rates, are required to prevent the a rules from being corrupted before the j3
rules have been completely learned. Furthermore, it is difficult to form long rule chains,
because each chunk further splits up the hidden unit population. Repeated splitting and
retraining of hidden units proved difficult, but the model did manage to learn an RI-R2R3 chunk that supersedes the RI-R2 chunk, so that ADE mutates directly to CFE. The
third rule was:
R3:
B~F/ C _E
"change B to F when between C and En
433
434
Touretzky and Elvgren
Output Change Pattern
Competitive
Rule Units
Input String Buffer
Input Change Pattern
(Training Only)
Figure 2: Architecture of the competitive learning chunker.
3
CHUNKING VIA COMPETITIVE LEARNING
Our second chunker, shown in Figure 2, minimizes interference between rules by using
competitive learning to assign each rule a dedicated unit. As in the previous case, the
model is taught its initial rules by showing it input buffer states and desired change buffer
states. Chunks are then formed by running strings through the input buffer and watching
for pairs of rules that fire sequentially. The model recruits new units for the chunks
and teaches them to produce the new change buffer patterns (formed by composing the
changes of the two original rules) in appropriate environments.
A number of technical problems had to be resolved in order to make this scheme work.
First, we want to assign a separate unit to each rule, but not to each training example;
otherwise the model will use too many units and not generalize well. Second, the
encoding for letters we chose (see Table 1) is based on a Cartesian product, and so input
patterns are highly overlapping and close together in Hamming space. This makes the
job of the competitive learning algorithm more difficult. Third, there must be some way
for chunks to take priority over the component rules from which they were fonned, so
that an input sequence like ADE fires the chunk RI-R2 rather than the original rule Rl.
As we trace through the operation of the chunker we will describe our solutions to these
problems.
Rule units in the competitive layer are in one of three states: inactive (waiting to be
recruited), plastic (currently undergoing learning), and active (weights finalized; ready to
compete and fire.) They also contain a simple integrator (a counter) that is used to move
them from the plastic to the active state. Initially all units are inactive and the counter
Rule Representations in a Connectionist Chunker
Table 1: Input code for both chunking models.
A
B
C
D
E
F
1
1
1
0
0
0
0
0
0
1
1
1
1
0
0
0
1
1
0
0
0
0
0
0
1
1
0
0
0
1
is zero. As in any competitive learning scheme, the rule units' input weights are kept
normalized to unit vectors (Rumelhart & Zipser, 1986).
When the teacher presents a novel instance, we must determine if there is already some
partially-trained rule unit whose weights should be shaped by this instance. Due to our
choice of input code, it is not possible to reliably assign training instances to rule units
based solely on the input pattern, because "similar" inputs (close in Hamming space)
may invoke entirely different rules. Our solution is to use the desired change pattern as
the primary index for selecting a pool of plastic rule units; the input buffer pattern is
then used as a secondary cue to select the most strongly activated unit from this pool.
Let's consider what happens with the training example DE - BE. The desired change
pattern "mutate segment 2 to a B" is fed to the competitive layer, and the network looks
for plastic rule units whose change patterns exactly match the desired pattern. 3 If no such
unit is found, one is allocated from the inactive pool, its status is changed to "plastic,"
its input buffer weights are set to match the pattern in the input buffer, and its change
pattern input and and change pattern output weig.hts are set according to the desired
change pattern.
Otherwise, if a pool of suitable plastic units already exists, the input pattern DE is
presented to the competitive layer and the selected plsatic units compete to see which
most closely matches the input The winning unit's input buffer weights are then adjusted
by competitive learning to move the weight vector slightly closer to this input buffer
vector. The unit's counter is also bumped.
Several presentations are normally required before a rule unit's input weights settle into
their correct values, since the unit must determine from experience which input bit values
are significant and which should be ignored. For example, rule S 1 in Table 2 (the asterisk
indicates a wildcard) can be learned from the training instances ACF and ADF, since as
Table 1 shows, the letters C and D in the second segment have no bits in common.
Therefore the learning algorithm will concentrate virtually all of the weight vector's
magnitude in the connections that specify "A" as the first segment and "F' as the third.
Each time a rule unit's weights are adjusted by competitive learning, its counter is in3The units' thresholds are raised so that they can only become active if their weight vectors match the input
change buffer vector exactly.
435
436
Touretzky and Elvgren
cremented. When the counter reaches a suitable value (currently 25), the unit switches
from the plastic to the active state. It is now ready to compete with other units for the
right to fire; its weights will not change further.
We now consider the formation of the model's first chunk. Assume that rules RI and
R2 have been acquired successfully. The model is trained by running random strings
through the input buffer and looking for sequences of rule firings. Suppose the model is
presented with the input string BFDADE. RI fires, producing BFDABE; this then causes
R2 to fire, producing BFDCBE. The model proceeds to form a chunk. The combined
change pattern specifies that the penultimate segment should be mutated to "B," and the
antepenultimate to "C." Since no plastic rule unit's change pattern weights match this
change, a fresh unit is allocated and its change buffer weights are set to reproduce this
pattern. The unit's input weights are set to detect the pattern BFDADE.
After several more examples of the RI-R2 firing sequence, the competitive learning
algorithm will discover that the first three input buffer positions can hold anything at all,
but the last three always hold ADE. Hence the weight vector will be concentrated on the
last three positions. When its counter reaches a value of 25, the rule unit will switch to
the active state.
Now consider the next time an input ending in ADE is presented. The network is in
performance mode now, so there is nothing in the input change buffer; the model is
looking only at the input string buffer. The RI unit will be fully satisfied by the input;
its normalized weight vector concentrates on just the last two positions, "DE," which
match exactly. The RI-R2 unit will also be fully satisfied; its normalized weight vector
looks for the sequence ADE. The latter unit is the one we want to win the competition.
We achieve this by scaling the activation function of competitive units by an additional
factor: the degree of distributedness of the weight vector. Units that distribute their input
weight over a larger number of connections likely represent complex chunks, and should
therefore have their activation boosted over rules with narrowly focused input vectors.
Once the unit encoding the RI-R2 chunk enters the active state, its more distributed input
weights assure that it will always win over the RI unit for an input like ADE. The RI
unit may still be useful to keep around, though, to handle a case like FDE -+ FBE that
does not trigger R2.
Sometimes a new chunk is learned that covers the same length input as the old, e.g.,
chunk RI-R2-R3 that maps ADE -+ CFE looks at exactly the same input positions as
chunk RI-R2. We therefore introduce one additional term into the activation function.
As part of the learning process, active units that contribute to the formation of a new
chunk are given a permanent, very small inhibitory bias. This ensures that RI-R2 will
always lose the competition to RI-R2-R3 once that chunk goes from plastic to active,
even though their weights are distributed to an equal degree.
Another special case that needs to be handled is when the competitive algorithm wrongly
splits a rule between two plastic units in the same pool, e.g., one unit might be assigned
the cases {A,B,C} ADE, and the other the cases {D.E,F} ADE. (In other words, one unit
looks for the bit pattern IOxxx in the first position, and the other unit looks for Olxxx.)
Rule Representations in a Connectionist Chunker
This is bad because it allows the weights of each unit to be more distributed than they need
to be. To correct the problem, whenever a plastic unit wins a competition our algorithm
makes sure that the nearest runner up is considerably less active than the winner. If its
activation is too high, the runner up is killed. This causes the survivor to readjust its
weights to describe the rule correctly, i.e., it will look for the input pattern ADE. If the
runner up was killed incorrectly (meaning it is really needed for some other rule), it will
be resurrected in response to future examples.
Finally, active units have a decay mechanism that is kept in check by the unit's firing
occasionally. If a unit does not fire for a long time (200 input presentations), its weights
decay to zero and it returns to the inactive state. This way. units representing chunks that
have been superseded will eventually be recycled.
4 DISCUSSION
Each of the two learning architectures has unique advantages. The backpropagation
learner can in principle learn arbitrarily complex rules. such as replacing a letter with
its successor. or reversing a subset of the input string. Its use of a distributed rule
representation allows knowledge of rule RI to participate in the forming of the RI-R2
chunk. However. this representation is also subject to interference effects. and as is often
the case with backprop. learning is slow.
The competitive architecture learns very quickly. It can form a greater number of chunks.
and can handle longer rule chains. since it avoids inteference by assigning a dedicated
unit to each new rule it learns.
Both learners are sensitive to changes in the distribution of input strings; new chunks
can form any time they are needed. Chunks that are no longer useful in the backprop
model will eventually fade away due to non-rehearsal; the hidden units that implement
these chunks will be recruited for other tasks. The competitive chunker uses a separate
decay mechanism to recycle chunks that have been superseded.
This work shows that connectionist techniques can yield novel and interesting solutions
to symbol processing problems. Our models are based on a sequence manipulation architecture that uses a symbolic description of the changes to be made (via the change
buffer), but the precise environments in which rules apply are never explicitly represented. Instead they are induced by the learning algorithm from examples of the models'
own behavior. Such self-supervised learning may play an important role in cognitive
development. Our work shows that it is possible to correctly chunk knowledge even
when one cannot predict the precise environment in which the chunks should apply.
Acknowledgements
This research was supported by a contract from Hughes Research Laboratories, by the
Office of Naval Research under contract number NOOOI4-86-K-0678. and by National
Science Foundation grant EET-8716324. We thank Allen Newell. Deirdre Wheeler. and
Akihiro Hirai for helpful discussions.
437
438
Touretzky and Elvgren
Table 2: Initial rule set for the competitive learning chunker.
SI:
S2:
S3:
S4:
S5:
A"'F
-+
BD
-+
{D,E,F}*E -+
{B,E}B
-+
{A,D}C -+
B*F
BF
{A,B,C}*A
CB
{C,F}C
Table 3: Chunks formed by the competitive learning chunker.
Chunk
EA*F -+ CB*F
ABD -+ CBF
AADF-+ CBFF
BE*E -+ CB*A
DEB -+ FEB
(Component Rules)
(SI,S4)
(SI,S2,S4)
(S I,S2,S I,S4)
(S3,S4)
(S4,S5)
Rererences
Laird, J. E., Newell, A., and Rosenbloom, P. S. (1987) Soar: An architecture for general
intelligence. Artificial Intelligence 33(1):1-64.
Newell, A. (1987) The 1987 William James Lectures: Unified Theories of Cognition.
Given at Harvard University.
Rurnelhart, D E., and Zipser, D. (1986) Feature discovery by competitive learning. In D.
E. Rumelhart and J. L. McClelland (eds.), Parallel Distributed Processing: Explorations
in the Microstructure oj Cognition. Cambridge, MA: MIT Press.
Touretzky. D. S. (1989a) Chunking in a connectionist network. Proceedings of the
Eleventh Annual Conference of the Cognitive Science Society, pp. 1-8. Hillsdale. NJ:
Erlbaum.
Touretzky, D. S. (1989b) Towards a connectionist phonology: the "many maps" approach to sequence manipulation. Proceedings of the Eleventh Annual Conference of the
Cognitive Science Society. pp. 188-195. Hillsdale. NJ: Erlbaurn.
Touretzky. D. S., and Wheeler. D. W. (1990) A computational basis for phonology. In D.
S. Touretzky (ed.), Advances in Neural Information Processing Systems 2. San Mateo.
CA: Morgan Kaufmann.
| 216 |@word version:2 retraining:1 bf:1 initial:4 contains:2 selecting:1 activation:4 assigning:1 si:3 must:4 bd:1 enables:1 motor:1 update:1 hts:1 cue:1 selected:1 intelligence:2 fbe:1 contribute:2 become:2 consists:1 combine:1 eleventh:2 behavioral:1 introduce:2 manner:1 acquired:1 indeed:1 behavior:4 integrator:1 detects:1 deirdre:1 becomes:1 discover:1 intennediate:1 what:2 string:13 minimizes:1 recruit:1 developed:1 unified:2 nj:2 exactly:4 unit:56 normally:1 grant:1 producing:2 before:2 installed:1 encoding:2 firing:6 solely:1 might:1 chose:1 mateo:1 unique:1 responsible:1 hughes:1 implement:2 writes:1 backpropagation:9 procedure:1 wheeler:3 word:1 symbolic:3 cannot:1 close:2 wrongly:1 context:1 writing:1 map:2 go:3 focused:1 splitting:1 fade:1 rule:76 examines:1 population:1 handle:2 enhanced:1 suppose:1 trigger:1 play:1 us:5 pa:1 assure:1 rumelhart:2 harvard:1 role:1 akihiro:1 enters:1 seg:4 ensures:1 counter:6 environment:5 insertion:1 trained:3 rewrite:2 segment:9 abd:1 learner:2 completely:1 basis:1 resolved:1 represented:3 derivation:3 describe:2 artificial:1 formation:4 whose:2 larger:2 otherwise:2 ability:1 mutates:1 laird:2 sequence:7 triggered:2 advantage:1 product:1 remainder:1 achieve:1 representational:1 description:1 competition:3 mutate:3 empty:1 produce:2 generating:1 incremental:2 nearest:1 school:1 job:1 implemented:1 differ:1 concentrate:2 bumped:1 closely:1 saved:1 correct:2 exploration:1 human:1 successor:1 settle:1 implementing:1 hillsdale:2 backprop:2 assign:3 microstructure:1 really:1 adjusted:3 hold:2 around:1 cb:4 cognition:3 predict:1 circuitry:3 lose:1 currently:2 sensitive:2 successfully:1 mit:1 always:3 rather:1 cbf:1 boosted:1 office:1 naval:1 improvement:1 chunked:6 indicates:1 survivor:1 check:1 contrast:1 detect:1 helpful:1 flaw:1 initially:1 hidden:8 reproduce:1 development:1 raised:1 special:1 equal:1 once:3 f3:1 shaped:1 never:1 placing:1 look:6 future:1 connectionist:11 preserve:1 national:1 fire:13 maintain:1 william:1 highly:1 runner:3 activated:1 chain:2 edge:2 closer:1 experience:1 perfonnance:1 old:1 desired:5 instance:4 cover:1 formulates:1 cost:1 subset:1 erlbaum:1 too:2 teacher:3 corrupted:2 accomplish:1 considerably:1 combined:2 chunk:37 contract:2 off:1 invoke:1 pool:5 together:1 quickly:1 again:1 satisfied:3 manage:1 priority:1 watching:1 cognitive:4 external:2 return:1 actively:1 distribute:1 de:3 permanent:1 caused:1 explicitly:1 ad:2 competitive:21 sort:1 start:1 parallel:1 mutation:1 formed:3 kaufmann:1 yield:1 generalize:1 mutated:1 plastic:11 history:1 explain:1 reach:2 touretzky:13 whenever:1 ed:2 rehearsal:3 pp:2 james:1 hamming:2 cur:2 proved:1 noooi4:1 knowledge:3 ea:1 back:1 adf:1 supervised:2 hirai:1 permitted:1 gillette:1 specify:1 response:1 though:2 strongly:1 furthermore:1 anywhere:1 stage:1 just:1 working:1 replacing:1 overlapping:1 mode:1 effect:1 contain:1 normalized:3 hence:1 assigned:1 read:1 laboratory:1 during:1 self:2 acquires:1 anything:1 dedicated:2 allen:1 meaning:1 novel:2 common:1 rl:2 winner:1 mellon:1 significant:1 s5:2 cambridge:1 ai:1 killed:2 had:3 cbe:2 longer:2 rosenbloom:2 feb:1 own:2 manipulation:4 occasionally:1 buffer:40 arbitrarily:1 morgan:1 additional:2 greater:1 accomplishes:1 determine:2 infer:1 technical:1 match:6 long:2 j3:4 represent:2 sometimes:1 want:2 source:1 allocated:2 sure:1 subject:1 recruited:2 virtually:1 induced:1 zipser:2 supersedes:1 counting:2 revealed:1 split:3 variety:1 switch:2 architecture:10 inactive:4 narrowly:1 handled:2 rurnelhart:1 cause:3 distributedness:1 ignored:1 useful:2 s4:6 concentrated:1 mcclelland:1 generate:1 specifies:1 exist:1 inhibitory:1 shifted:1 s3:2 track:1 correctly:3 carnegie:1 write:1 waiting:1 taught:1 threshold:1 achieving:1 prevent:1 prey:2 kept:2 counteract:1 compete:3 letter:3 scaling:1 bit:3 entirely:1 layer:10 followed:2 annual:2 occur:1 deb:1 ri:28 according:1 combination:1 request:2 recycle:1 chunker:13 describes:1 slightly:1 happens:1 explained:1 interference:4 chunking:11 previously:1 eventually:3 mechanism:3 r3:3 needed:2 fed:1 operation:1 apply:3 away:1 appropriate:1 original:5 running:2 phonology:2 readjust:1 classical:1 society:2 move:2 already:2 primary:2 exhibit:1 win:3 separate:3 thank:1 penultimate:1 participate:1 fresh:1 code:2 length:1 index:1 difficult:3 teach:2 trace:2 reliably:1 allowing:1 incorrectly:1 looking:2 precise:2 abe:2 david:1 pair:1 required:2 specified:1 connection:7 learned:4 deletion:1 temporary:1 fde:1 proceeds:1 pattern:25 program:1 oj:1 memory:1 shifting:3 ia:1 suitable:2 force:1 representing:1 scheme:2 superseded:2 ready:2 acknowledgement:1 discovery:1 contributing:1 fully:2 lecture:1 interesting:1 limitation:1 asterisk:1 foundation:1 degree:2 principle:1 share:1 production:1 changed:1 supported:2 cfe:2 last:3 bias:1 wide:1 distributed:6 ending:1 avoids:1 made:2 fonned:1 san:1 finalized:1 status:1 eet:1 keep:1 active:11 sequentially:1 pittsburgh:1 continuous:1 table:6 learn:3 robust:1 fragility:1 composing:1 ca:1 recycled:1 requested:1 complex:3 did:1 s2:3 nothing:1 repeated:1 en:1 slow:1 position:8 acf:1 elvgren:5 winning:1 perceptual:1 ib:1 third:3 learns:2 bad:1 showing:1 symbol:3 r2:25 undergoing:1 decay:3 derives:1 exists:1 magnitude:1 cartesian:1 tc:1 likely:1 forming:1 soar:3 partially:1 applies:3 newell:5 ma:1 presentation:3 towards:1 replace:1 change:46 reversing:1 ade:15 secondary:1 wildcard:1 rererences:1 select:1 latter:1 |
1,274 | 2,160 | On the Dirichlet Prior and Bayesian
Regularization
Harald Steck
Artificial Intelligence Laboratory
Massachusetts Institute of Technology
Cambridge, MA 02139
[email protected]
Tommi S. Jaakkola
Artificial Intelligence Laboratory
Massachusetts Institute of Technology
Cambridge, MA 02139
[email protected]
Abstract
A common objective in learning a model from data is to recover
its network structure, while the model parameters are of minor interest. For example, we may wish to recover regulatory networks
from high-throughput data sources. In this paper we examine how
Bayesian regularization using a product of independent Dirichlet
priors over the model parameters affects the learned model structure in a domain with discrete variables. We show that a small
scale parameter - often interpreted as "equivalent sample size" or
"prior strength" - leads to a strong regularization of the model
structure (sparse graph) given a sufficiently large data set. In particular, the empty graph is obtained in the limit of a vanishing scale
parameter. This is diametrically opposite to what one may expect
in this limit, namely the complete graph from an (unregularized)
maximum likelihood estimate. Since the prior affects the parameters as expected, the scale parameter balances a trade-off between
regularizing the parameters vs. the structure of the model. We
demonstrate the benefits of optimizing this trade-off in the sense
of predictive accuracy.
1
Introduction
Regularization is essential when learning from finite data sets. In the Bayesian approach, regularization is achieved by specifying a prior distribution over the parameters and subsequently averaging over the posterior distribution. This regularization
provides not only smoother estimates of the parameters compared to maximum
likelihood but also guides the selection of model structures.
It was pointed out in [6] that a very large scale parameter of the Dirichlet prior can
degrade predictive accuracy due to severe regularization of the parameter estimates.
We complement this discussion here and show that a very small scale parameter
can lead to poor over-regularized structures when a product of (conjugate) Dirichlet
priors is used over multinomial conditional distributions (Section 3). Section 4
demonstrates the effect of the scale parameter and how it can be calibrated. We
focus on the class of Bayesian network models throughout this paper.
2
Regularization of Parameters
We briefly review Bayesian regularization of parameters. We follow the assumptions outlined in [6] : multinomial sample, complete data, parameter modularity,
parameter independence, and Dirichlet prior. Note that the Dirichlet prior over
the parameters is often used for two reasons: (1) the conjugate prior permits analytical calculations, and (2) the Dirichlet prior is intimately tied to the desirable
likelihood-equivalence property of network structures [6]. The Dirichlet prior over
the parameters 8' I11"i is given by
(1)
where 8Xi l11"i pertains to variable X i in state Xi given that its parents IIi are in joint
state 'Tri . The number of variables in the domain is denoted by n, and i = 1, ... , n.
The normalization terms in Eq. 1 involve the Gamma function r(?). There are
a number of approaches to specifying the positive hyper-parameters O:Xi ,11"i of the
Dirichlet prior [2, 1, 6] ; we adopt the common choice,
(2)
where p is a (marginal) prior distribution over the (joint) states, as this assignment
ensures likelihood equivalence of the network structures [6]. Due to lack of prior
knowledge, p is often chosen to be uniform, p(Xi,'Tri ) = 1/ (IXil?IIIil), where lXii ,
IIIi l denote the number of (joint) states [1]. The scale parameter 0: of the Dirichlet
prior is positive and independent of i, i.e. , 0: = L Xi ,11"i O:Xi ,11"i '
The average parameter value OXi l11" i ' which typically serves as the regularized parameter estimate given a network structure m , is given by
o
= E p(Ox i l ~i I D,m) [ 8
]=
Xi l11"i
Xi l11"i -
+
i, 11"i
+ O:X
'
Q
N Xi ,11"i
N
7ri
(3)
7ri
where N Xi ,11"i are the cell-counts from data D; E[?] is the expectation. Positive
hyper-parameters O:X i, 11"i lead to regularized parameter estimates, i.e., the estimated
parameters become "smoother" or " less extreme" when the prior distribution p is
close to uniform. An increasing scale parameter 0: leads to a stronger regularization, while in the limit 0: -+ 0, the (unregularized) maximum likelihood estimate is
obtained, as expected.
3
Regularization of Structure
In the remainder of this paper, we outline effects due to Bayesian regularization of
the Bayesian network structure when using a product of Dirichlet priors. Let us
briefly introduce relevant notation.
In the Bayesian approach to structure learning, the posterior probability of the
network structure m is given by p(mID) = p(Dlm)p(m)/p(D), where p(D) is the
(unknown) probability of given data D , and p(m) denotes the prior distribution over
the network structures; we assume p(m) > 0 for all m. Following the assumptions
outlined in [6], including the Dirichlet prior over the parameters 8, the marginal
likelihood p(Dlm) = Ep(O lm) [p(Dlm , 8)] can be calculated analytically. Pretending
that the (i.i.d.) data arrived in a sequential manner , it can be written as
N
p(Dlm) =
II II
n
N(k-l)
k = l i=l
N 7r ik
+ 0:
X:(':~l)
k
k
Xi ,11"i ,
+ O:11"k
i
(4)
where N(k-l) denotes the counts implied by data D(k-l) seen before step k along
the sequence (k = 1, ... , N). The (joint) state of variable Xi and its parents IIi
occurring in the kth data point is denoted by xf, 7rf. In Eq. 4, we also decomposed
the joint probability into a product of conditional probabilities according to the
Bayesian network structure m. Eq. 4 is independent of the sequential ordering of
the data points, and the ratio in Eq. 3 corresponds to the one in Eq. 4 when based
on data D(k - l) at each step k along the sequence.
3.1 Limit of Vanishing Scale-Parameter
This section is concerned with the limit of a vanishing scale parameter of the Dirichlet prior, a -+ O. In this limit Bayesian regularization depends crucially on the
number of zero-cell-counts in the contingency table implied by the data, or in other
words, on the number of different configurations (data points) contained in the
data. Let the Effective Number of Parameters (EP) be defined as
n
dk';) =
l: [l: I(Nxi,1rJ - l: I(N1rJ ],
(5)
where N Xi ,1ri' N1ri are the (marginal) cell counts in the contingency table implied
by data D; m refers to the Bayesian network structure, and 1(?) is an indicator
function such that I(z) = 0 if z = 0 and I(z) = 1 otherwise. When all cell
counts are positive, EP is identical to the well-known number of parameters (P),
dk';) = m ) = L:i(IXil - l)IIIil, which play an important role in regularizing the
learned network structure. The key difference is that EP accounts for zero-cellcounts implied by the data.
Let us now consider the behavior of the marginal likelihood (cf. Eq. 4) in the limit
of a small scale parameter a. We find
Proposition 1: Under the assumptions concerning the prior distribution outlined
in Section 2, the marginal likelihood of a Bayesian network structure m vanishes in
the limit a -+ 0 if the data D contain two or more different configurations. This
property is independent of the network structure. The leading polynomial order is
given by
d(=l
p(Dlm) "-' a EP
as a -+ 0,
(6)
4
which depends both on the network structure and the data. However, the dependence
on the data is through the number of different data points only. This holds independently of a particular choice of strictly positive prior distributions P(Xi ' IIi). If
the prior over the network structures is strictly positive, this limiting behavior also
holds for the posterior probability p( miD) .
In the following we give a derivation of Proposition 1 that also facilitates the intuitive
understanding of the result. First, let us consider the behavior of the Dirichlet
distribution in the limit a -+ O. The hyper-parameters a X i , 1r i vanish when a -+ 0,
and thus the Dirichlet prior converges to a discrete distribution over the parameter
simplex in the sense that the probability mass concentrates at a particular, randomly
chosen corner of the simplex containing B. I1ri (cf. [9]). Since the randomly chosen
points (for different 7ri, i) do not change when sampling (several) data points from
the distribution implied by the model , it follows immediately that the marginal
likelihood of any network structure vanishes whenever there are two or more different
configurations contained in the data.
This well-known fact also shows that the limit a -+ 0 actually corresponds to a very
strong prior belief [9, 12]. This is in contrast to many traditional interpretations
where the limit a -+ 0 is considered as "no prior information", often motivated
by Eq. 3. As pointed out in [9, 12], the interpretation of the scale parameter a
as "equivalent sample size" or as the" strength" of prior belief may be misleading,
particularly in the case where O:X i, 1ri < 1 for some configurations Xi, 7ri. A review
of different notions of "noninformative" priors (including their limitations) can be
found in [7]. Note that the noninformative prior in the sense of entropy is achieved
by setting O:Xi , 1ri = 1 for each Xi, 7ri and for all i = 1, ... , n. This is the assignment
originally proposed in [2]; however , this assignment generally is inconsistent with
Eq. 2, and hence with likelihood equivalence [6].
In order to explain the behavior of the marginal likelihood in leading order of the
scale parameter 0:, the properties of the Dirichlet distribution are not sufficient by
themselves. Additionally, it is essential that the probability distribution described
by a Bayesian network decomposes into a product of conditional probabilities, and
that there is a Dirichlet prior pertaining to each variable for each parent configuration. All these Dirichlet priors are independent of each other under the standard
assumption of parameter independence. Obviously, the ratio (for given k and i) in
Eq. 4 can only vanish in the limit 0: --+ 0 if N(~ - ~ = 0 while N(~- l) > 0; in other
Xi , 7r i
7ri
words, the parent-configuration 7rf must already have occurred previously along
the sequence (7rf is "old"), while the child-state xf occurs simultaneously with this
parent-state for the first time (xf, 7rf is "new"). In this case, the leading polynomial order of the ratio (for given k and i) is linear in 0:, assuming P(Xi' IIi) > 0;
otherwise the ratio (for given k and i) converges to a finite positive value in the
limit 0: --+ O. Consequently, the dependence of the marginal likelihood in leading
polynomial order on 0: is completely determined by the number of different configurations in the data. It follows immediately that the leading polynomial order in
0: is given by EP (d. Eq. 5). This is because the first term counts the number of
all the different joint configurations of Xi , IIi in the data, while the second term
ensures that EP counts only those configurations where (xf, 7rf) is " new" while 7rf
is "old".
Note that the behavior of the marginal likelihood in Proposition 1 is not entirely
determined by the network structure in the limit 0: --+ 0, as it still depends on the
data. This is illustrated in the following example. First, let us consider two binary
variables, Xo and Xl, and the data D containing only two data points, say (0,0)
and (1,1). Given data D, three Dirichlet priors are relevant regarding graph ml,
Xo --+ Xl, but only two Dirichlet priors pertain to the empty graph, mo. The
resulting additional "flexibility" due to an increased number of priors favours more
complex models: p(Dlmd ~ 0:, while p(Dlmo) ~ 0: 2 . Second, let us now assume
that all possible configurations occur in data D. Then we still have p(Dlmo) ~ 0: 2
for the empty graph. Concerning graph ml, however , the marginal likelihood now
also involves the vanishing terms due to the two priors pertaining to BX1 lxo =o and
BXl lxo=l, and it hence becomes p(Dlmd ~ 0: 3 .
This dependence on the data can be formalized as follows. Let us compare the
marginal likelihoods of two graphs, say m+ and m - . In particular, let us consider
two graphs that are identical except for a single edge, say A +- B between the
variables A and B. Let the edge be present in graph m+ and absent in m-. The
fact that the marginal likelihood decomposes into terms pertaining to each of the
variables (d. Eq. 4) entails that all the terms regarding the remaining variables
cancel out in the Bayes factor p(Dlm+)/p(Dlm-), which is the standard relative
Bayesian score. With the definition of the Effective Degrees of Freedom (EDF)l
(7)
we immediately obtain from Proposition 1 that p(Dlm+)/p(Dlm- ) ~
INote that EDF are not necessarily non-negative.
o:dEDF
in the
limit a -+ 0, and hence
Proposition 2: Let m+ and m- be the two network structures as defined above.
Let the prior belief be given according to Eq. 2. Then in the limit a -+ 0:
{-oo
I p(Dlm+)
if d EDF > 0,
( )
og p(Dlm- ) -+
+00 if dEDF < O.
8
The result holds independently of a particular choice of strictly positive prior distributions P(Xi' IIi). If the prior over the network structures is strictly positive, this
limiting behavior also holds for the posterior ratio.
A positive value of the log Bayes factor indicates that the presence of the edge
A f- B is favored , given the parents IIA ; conversely, a negative relative score
suggests that the absence of this edge is preferred. The divergence of this relative
Bayesian score implies that there exists a (small) positive threshold value ao > 0
such that, for any a < ao, the same graph(s) are favored as in the limit.
Since Proposition 2 applies to every edge in the network, it follows immediately
that the empty (complete) graph is assigned the highest relative Bayesian score
when EDF are positive (negative). Regularization of network structure in the case
of positive EDF is therefore extreme, permitting only the empty graph. This is
precisely the opposite of what one may have expected in this limit, namely the
complete graph corresponding to the unregularized maximum likelihood estimate
(MLE). In contrast, when EDF are negative, the complete graph is favored. This
agrees with MLE.
Roughly speaking, positive (negative) EDF correspond to large (small) data sets.
It is thus surprising that a small data set, where one might expect an increased
restriction on model complexity, actually gives rise to the complete graph, while
a large data set yields the - most regularized - empty graph in the limit a -+ O.
Moreover, it is conceivable that a "medium" sized data set may give rise to both
positive and negative EDF. This is because the marginal contingency tables implied
by the data with respect to a sparse (dense) graph may contain a small (large)
number of zero-cell-counts. The relative Bayesian score can hence become rather
unstable in this case, as completely different graph structures are optimal in the
limit a -+ 0, namely graphs where each variable has either the maximal number of
parents or none.
Note that there are two reasons for the hyper-parameters a Xi , i to take on small
values (cf. Eq. 2): (1) a small equivalent sample size a, or (2) a large number of
joint states, i.e. IXi l? IIIil ? a , due to a large number of parents (with a large
number of states). Thus, these hyper-parameters can also vanish in the limit of a
large number of configurations (x , 1f) even though the scale parameter a remains
fixed. This is precisely the limit defining Dirichlet processes [4], which, analogously,
produce discrete samples. With a finite data set and a large number of joint configurations, only the typical limit in Proposition 2 is possible. This follows from the
fact that a large number of zero-cell-counts forces EDF to be negative. The surprising behavior implied by Proposition 2 therefore does not carryover to Dirichlet
processes. As found in [8], however, the use of a product of Dirichlet process priors
in non parametric inference can also lead to surprising effects.
When dEDF = 0, it is indeed true that the value of the log Bayes factor can converge
to any (possibly finite) value as a -+ O. Its value is determined by the priors
P(Xi ' IIi), as well as by the counts implied by the data. The value of the Bayes
factor can be therefore easily set by adjusting the prior weights p(Xi' 1fi).
1f
3.2
Large Scale-Parameter
In the other limiting case, where a -+ 00 , the Bayes factor approaches a finite value,
which in general depends on the given data and on the prior distributions p(Xi' IIi).
lBF
2
1.5
1
0.5
-0.5
-1
z=3
z=o
~
100
150
200
250
300
a
Figure 1: The log Bayes factor (lBF) is depicted as a function of the scale parameter
It is assumed that the two variables A and B are binary and have no parents;
and that the "data" imply the contingency table: NA = O,B= O = NA = l,B = l = 10 + z
and NA=l,B=O = NA=O,B=l = 10 - z, where z is a free parameter determining the
statistical dependence between A and B. The prior p(Xi,II i ) was chosen to be
uniform.
0:.
This can be seen easily by applying the Stirling approximation in the limit 0: -+ 00
after rewriting Eq. 4 in terms of Gamma functions (cf. also [2, 6]). When the
popular choice of a uniform prior p(Xi,II i ) is used [1], then
p(Dlm+)
log p(Dlm-) -+ 0
as
0:-+00,
(9)
which is independent of the data. Hence, neither the presence nor the absence of
the edge between A and B is favored in this limit. Given a uniform prior over the
network structures, p(m) =const, the posterior distribution p(mID) over the graphs
thus becomes increasingly spread out as 0: grows, permitting more variable network
structures.
The behavior of the Bayes factor between the two limits 0: -+ 0 and 0: -+ 00 is exemplified for positive EDF in Figure 1: there are two qualitatively different behaviors,
depending on the degree of statistical dependence between A and B. A sufficiently
weak dependence results in a monotonically increasing Bayes factor which favors the
absence of the edge A +- B at any finite value of 0:. In contrast, given a sufficiently
strong dependence between A and B, the log Bayes factor takes on positive values
for all (finite) 0: exceeding a certain value 0:+ of the scale parameter. Moreover, 0:+
grows as the statistical dependence between A and B diminishes. Consequently,
given a domain with a range of degrees of statistical dependences, the number of
edges in the learned graph increases monotonically with growing scale parameter 0:
when each variable has at most one parent (i. e., in the class of trees or forests). This
is because increasingly weaker statistical dependencies between variables are recovered as 0: grows; the restriction to forests excludes possible "interactions" among
(several) parents of a variable. As suggested by our experiments, this increase in
the number of edges can also be expected to hold for general Bayesian network
structures (although not necessarily in a monotonic way).
This reveals that regularization of network structure tends to diminish with a growing scale parameter. Note that this is in the opposite direction to the regularization
of parameters (cf. Section 2). Hence, the scale parameter 0: of the Dirichlet prior
determines the trade-off between regularizing the parameters vs. the structure of the
Bayesian network model.
If a uniform prior over the network structures is chosen, p(m) = const, the above
discussion also holds for the posterior ratio (instead of the Bayes factor). The
behavior is more complicated, however, when a non-uniform prior is assumed. For
instance, when a prior is chosen that penalizes the presence of edges, the posterior
favours the absence of an edge not only when the scale parameter is sufficiently
small, but also when it is sufficiently large. This is apparent from Fig. 1, when the
log Bayes factor is compared to a positive threshold value (instead of zero).
4
Example
This section exemplifies that the entire model (parameters and structure) has to
be considered when learning from data. This is because regularization of model
structure diminishes, while regularization of parameters increases with a growing
scale parameter a of the Dirichlet prior, as discussed in the previous sections.
When the entire model is taken into account, one can use a sensitivity analysis in
order to determine the dependence of the learned model on the scale parameter
a, given the prior p(Xi' IIi) (cf. Eq. 2). The influence of the scale parameter a
on predictive accuracy of the model can be assessed by cross-validation or, in a
Bayesian approach, prequential validation [11, 3]. Another possibility is to treat
the scale parameter a as an additional parameter of the model to be learned from
data. Hence, prior belief regarding the parameters e can then enter only through
the (normalized) distributions P(Xi' IIi). Howeverl. note that this is sufficient to
determine the (average) prior parameter estimate (cf. Eq. 3) , i.e., when N = O.
Assuming an (improper) uniform prior distribution over a, its posterior distribution
is p(aID) ex: p(Dla), given data D. Then aD = argmaxaP(Dla), where p(Dla) =
I: m P(Dla,m)p(m)2 can be calculated exactly if the summation is feasible (like in
the example below). Alternatively, assuming that the posterior over a is strongly
peaked, the likelihood may also be approximated by summing over the k most likely
graphs m only (k = 1 in the most extreme case; empirical Bayes). Subsequently,
model structure m and parameters B can be learned with respect to the Bayesian
score employing aD.
e
In the following, the effect of various values assigned to the scale parameter a is
exemplified concerning the data set gathered from Wisconsin high-school students
by Sewell and Shah [10]. This domain comprises 5 discrete variables, each with
2 or 4 states; the sample size is 10,318. In this small domain, exhaustive search
in the space of Bayesian network structures is feasible (29,281 graphs). Both the
prior distributions p(m) for all m and P(Xi' IIi) are chosen to be uniform. Figure 2
shows that the number of edges in the graph with the highest posterior probability
grows with an increasing value of the scale parameter, as expected (cf. Section 3).
In addition, cross-validation indicates best predictive accuracy of the learned model
at a ~ 100, ... ,300, while the likelihood p(Dla) takes on its maximum at aD ~ 69.
Both approaches agree on the same network structure, which is depicted in Fig. 3.
This graph can easily be interpreted in a causal manner, as outlined in [5].3 We
note that this graph was also obtained in [5] due, however , to additional constraints
concerning network structure, as a rather small prior strength of a = 5 was used. For
comparison, Fig. 3 also shows the highest-scoring unconstraint graph due to a = 5,
which does not permit a causal interpretation, cf. also [5]. This illustrates that
the "right" choice of the scale parameter a of the Dirichlet prior, when accounting
for both model structure and parameters, can have a crucial impact on the learned
network structure and the resulting insight in the ("true") dependencies among the
variables in the domain.
2We assume that m and a are independent a priori, p(mla) = p(m).
3Since we did not impose any constraints on the network structure, unlike to [5] ,
Markov-equivalence leaves the orientation of the edge between the variables IQ and CP
unspecified.
a
5
50
100
200
300
500
1, 000
a.
6
7
7
7
7
7
8
XV 5
0.045
0.044
0.040
0.040
0.040
0.042
0.047
p(D la)
p(D laD)
10 ?w
0.13
0.05
10- 14
10- 30
10- 65
10- 151
Figure 2: As a function of a: number of
arcs (a.) in the highest-scoring graph;
average KL divergence in 5-fold crossvalidation (XV 5), std= 0.006; likelihood
of a when treated as an additional model
parameter (aD = 69).
SES: socioeconomic status SEX: gender of student
PE: parental encouragement CP: college plans
IQ: intelligence quotient
Figure 3: Highest-scoring (unconstraint)
graphs when a = 5 (left) , and when a =
46, ... ,522 (right). Note that the latter
graph can also be obtained at a = 5 when
additional constraints are imposed on the
structure, cf. [5].
Acknowledgments
We would like to thank Chen-Hsiang Yeang and the anonymous referees for valuable
comments. Harald Steck acknowledges support from the German Research Foundation (DFG) under grant STE 1045/1-1. Tommi Jaakkola acknowledges support
from Nippon Telegraph and Telephone Corporation, NSF ITR grant IIS-0085836,
and from the Sloan Foundation in the form of the Sloan Research Fellowship.
References
[1] W. Buntine. Theory refinement on Bayesian networks. Conference on Uncertainty in Artificial Intellig ence, pages 52- 60. Morgan Kaufmann, 1991.
[2] G. Cooper and E. Herskovits. A Bayesian method for the induction of probabilistic networks from data. Machine Learning, 9:309- 47, 1992.
[3] A. P. Dawid. Statistical theory. The prequential approach. Journal of the Royal
Statistical Society, Series A , 147:277- 305, 1984.
[4] T. S. Ferguson. A Bayesian analysis of some nonparametric problems. Annals
of Statistics, 1:209- 30, 1973.
[5] D. Heckerman. A tutorial on learning with Bayesian networks. In M. I. Jordan
(Ed.), Learning in Graphical Models, pages 301- 54. Kluwer , 1996.
[6] D. Heckerman, D. Geiger, and D. M. Chickering. Learning Bayesian networks:
the combination of knowledge and statistical data. Machine Learning, 20:197243,1995.
[7] R. E . Kass and L. Wasserman. Formal rules for selecting prior distributions:
a review and annotated bibliography. Technical Report 583, CMU, 1993.
[8] S. Petrone and A. E. Raftery. A note on the Dirichlet process prior in Bayesian
nonparametric inference with partial exchangeability. Technical Report 297,
University of Washington, Seattle, 1995.
[9] J. Sethuraman and R. C. Tiwari. Convergence of Dirichlet measures and the
interpretation of their parameter. In S. S. Gupta and J. O. Berger (Eds.),
Statistical Decision Theory and Related Topics III, pages 305- 15, 1982.
[10] W. Sewell and V. Shah. Social class, parental encouragement, and educational
aspirations. American Journal of Sociology, 73:559- 72, 1968.
[11] M. Stone. Cross-validatory choice and assessment of statistical predictions.
Journal of the Royal Statistical Society, Series B, 36:111- 47, 1974.
[12] S. G. Walker and B. K. Mallick. A note on the scale parameter of the Dirichlet
process. The Canadian Journal of Statistics, 25:473- 9, 1997.
| 2160 |@word briefly:2 polynomial:4 stronger:1 sex:1 steck:2 diametrically:1 crucially:1 accounting:1 configuration:12 series:2 score:6 selecting:1 recovered:1 ka:1 surprising:3 written:1 must:1 ixil:2 noninformative:2 v:2 intelligence:3 leaf:1 vanishing:4 provides:1 along:3 mla:1 become:2 ik:1 manner:2 introduce:1 indeed:1 expected:5 roughly:1 themselves:1 examine:1 nor:1 growing:3 behavior:10 decomposed:1 increasing:3 becomes:2 notation:1 moreover:2 mass:1 medium:1 what:2 interpreted:2 unspecified:1 corporation:1 every:1 exactly:1 demonstrates:1 grant:2 positive:18 before:1 treat:1 tends:1 limit:26 xv:2 might:1 equivalence:4 specifying:2 conversely:1 suggests:1 range:1 acknowledgment:1 empirical:1 dla:5 word:2 refers:1 close:1 selection:1 pertain:1 nxi:1 applying:1 influence:1 restriction:2 equivalent:3 imposed:1 educational:1 independently:2 formalized:1 immediately:4 wasserman:1 insight:1 rule:1 notion:1 limiting:3 annals:1 play:1 ixi:1 nippon:1 referee:1 dawid:1 approximated:1 particularly:1 std:1 ep:7 role:1 ensures:2 improper:1 ordering:1 trade:3 highest:5 valuable:1 prequential:2 vanishes:2 complexity:1 aspiration:1 bx1:1 predictive:4 completely:2 easily:3 joint:8 various:1 derivation:1 effective:2 pertaining:3 artificial:3 hyper:5 exhaustive:1 apparent:1 say:3 s:1 otherwise:2 favor:1 statistic:2 obviously:1 sequence:3 analytical:1 interaction:1 product:6 maximal:1 remainder:1 ste:1 relevant:2 flexibility:1 intuitive:1 crossvalidation:1 seattle:1 parent:11 empty:6 convergence:1 produce:1 converges:2 oo:1 depending:1 iq:2 school:1 minor:1 eq:16 strong:3 quotient:1 involves:1 implies:1 tommi:3 concentrate:1 direction:1 annotated:1 subsequently:2 ao:2 anonymous:1 proposition:8 summation:1 strictly:4 hold:6 sufficiently:5 considered:2 diminish:1 mo:1 lm:1 adopt:1 diminishes:2 agrees:1 mit:2 rather:2 exchangeability:1 og:1 jaakkola:2 exemplifies:1 focus:1 likelihood:20 indicates:2 contrast:3 sense:3 inference:2 ferguson:1 typically:1 entire:2 among:2 orientation:1 denoted:2 favored:4 priori:1 plan:1 marginal:13 validatory:1 washington:1 sampling:1 identical:2 cancel:1 throughput:1 peaked:1 simplex:2 carryover:1 report:2 randomly:2 gamma:2 simultaneously:1 divergence:2 dfg:1 freedom:1 interest:1 possibility:1 severe:1 extreme:3 edge:13 partial:1 tree:1 old:2 penalizes:1 causal:2 sociology:1 increased:2 instance:1 ence:1 stirling:1 assignment:3 uniform:9 socioeconomic:1 buntine:1 dependency:2 calibrated:1 sensitivity:1 probabilistic:1 off:3 telegraph:1 analogously:1 na:4 containing:2 possibly:1 corner:1 sewell:2 american:1 leading:5 account:2 student:2 sloan:2 depends:4 ad:4 recover:2 bayes:12 complicated:1 accuracy:4 kaufmann:1 correspond:1 yield:1 gathered:1 weak:1 bayesian:28 none:1 explain:1 whenever:1 ed:2 definition:1 adjusting:1 massachusetts:2 popular:1 knowledge:2 tiwari:1 actually:2 originally:1 follow:1 ox:1 though:1 strongly:1 assessment:1 lack:1 grows:4 effect:4 contain:2 true:2 normalized:1 regularization:18 analytically:1 hence:7 assigned:2 laboratory:2 illustrated:1 stone:1 arrived:1 outline:1 complete:6 demonstrate:1 cp:2 fi:1 common:2 multinomial:2 discussed:1 interpretation:4 occurred:1 kluwer:1 lad:1 cambridge:2 ai:2 enter:1 encouragement:2 outlined:4 iia:1 pointed:2 entail:1 posterior:10 optimizing:1 certain:1 binary:2 scoring:3 seen:2 morgan:1 additional:5 impose:1 converge:1 determine:2 monotonically:2 ii:5 smoother:2 desirable:1 rj:1 technical:2 xf:4 calculation:1 cross:3 lxo:2 concerning:4 permitting:2 mle:2 impact:1 prediction:1 expectation:1 cmu:1 normalization:1 harald:3 achieved:2 cell:6 addition:1 fellowship:1 iiii:1 walker:1 source:1 crucial:1 unlike:1 tri:2 comment:1 facilitates:1 inconsistent:1 jordan:1 presence:3 canadian:1 iii:12 concerned:1 affect:2 independence:2 opposite:3 regarding:3 itr:1 absent:1 favour:2 motivated:1 speaking:1 generally:1 involve:1 nonparametric:2 mid:3 nsf:1 herskovits:1 tutorial:1 estimated:1 discrete:4 key:1 threshold:2 neither:1 rewriting:1 graph:31 excludes:1 uncertainty:1 throughout:1 geiger:1 decision:1 entirely:1 fold:1 strength:3 occur:1 precisely:2 constraint:3 ri:9 bibliography:1 according:2 combination:1 poor:1 conjugate:2 heckerman:2 increasingly:2 intimately:1 inote:1 xo:2 dlm:13 yeang:1 taken:1 unregularized:3 intellig:1 agree:1 previously:1 remains:1 count:10 german:1 serf:1 oxi:1 permit:2 shah:2 denotes:2 dirichlet:29 cf:10 remaining:1 graphical:1 const:2 society:2 implied:8 objective:1 already:1 occurs:1 parametric:1 dependence:10 traditional:1 conceivable:1 kth:1 thank:1 degrade:1 topic:1 unstable:1 reason:2 induction:1 assuming:3 berger:1 ratio:6 balance:1 negative:7 rise:2 unknown:1 pretending:1 l11:4 markov:1 arc:1 finite:7 defining:1 complement:1 namely:3 kl:1 learned:8 suggested:1 parental:2 below:1 exemplified:2 rf:6 including:2 royal:2 belief:4 mallick:1 treated:1 force:1 regularized:4 indicator:1 technology:2 misleading:1 imply:1 sethuraman:1 raftery:1 acknowledges:2 prior:60 review:3 understanding:1 determining:1 relative:5 wisconsin:1 expect:2 limitation:1 validation:3 foundation:2 contingency:4 degree:3 sufficient:2 edf:10 free:1 guide:1 weaker:1 formal:1 institute:2 sparse:2 benefit:1 calculated:2 qualitatively:1 refinement:1 employing:1 social:1 preferred:1 status:1 ml:2 reveals:1 summing:1 assumed:2 xi:30 alternatively:1 search:1 regulatory:1 modularity:1 decomposes:2 table:4 additionally:1 forest:2 complex:1 necessarily:2 domain:6 did:1 dense:1 spread:1 child:1 fig:3 hsiang:1 aid:1 cooper:1 comprises:1 wish:1 exceeding:1 xl:2 tied:1 vanish:3 pe:1 chickering:1 dk:2 gupta:1 essential:2 exists:1 i11:1 sequential:2 illustrates:1 occurring:1 chen:1 entropy:1 depicted:2 likely:1 contained:2 applies:1 monotonic:1 gender:1 corresponds:2 determines:1 ma:2 conditional:3 sized:1 consequently:2 absence:4 feasible:2 change:1 determined:3 except:1 typical:1 telephone:1 averaging:1 la:1 college:1 support:2 latter:1 assessed:1 pertains:1 lbf:2 regularizing:3 ex:1 |
1,275 | 2,161 | Analysis of Information in Speech based
on MANOVA
Sachin s. Kajarekarl and Hynek Hermansky l ,2
1 Department of Electrical and Computer Engineering
OGI School of Science and Engineering at OHSU
Beaverton, OR
2International Computer Science Institute
Berkeley, CA
{ sachin,hynek} @asp.ogi.edu
Abstract
We propose analysis of information in speech using three sources
- language (phone), speaker and channeL Information in speech is
measured as mutual information between the source and the set of
features extracted from speech signaL We assume that distribution of features can be modeled using Gaussian distribution. The
mutual information is computed using the results of analysis of
variability in speech. We observe similarity in the results of phone
variability and phone information, and show that the results of the
proposed analysis have more meaningful interpretations than the
analysis of variability.
1
Introduction
Speech signal carries information about the linguistic message, the speaker, the
communication channeL In the previous work [1, 2], we proposed analysis of information in speech as analysis of variability in a set of features extracted from
the speech signal. The variability was measured as covariance of the features , and
analysis was performed using using multivariate analysis of variance (MANOVA).
Total variability was divided into three types of variabilities, namely, intra-phone
(or phone) variability, speaker variability, and channel variability. Effect of each
type was measured as its contribution to the total variability.
In this paper, we extend our previous work by proposing an information-theoretic
analysis of information in speech. Similar to MANOVA, we assume that speech
carries information from three main sources- language, speaker, and channeL We
measure information from a source as mutual information (MI) [3] between the
corresponding class labels and features. For example, linguistic information is measured as MI between phone labels and features. The effect of sources is measured
in nats (or bits). In this work, we show it is easier to interpret the results of this
analysis than the analysis of variability.
In general, MI between two random variables X and Y can be measured using
three different methods [4]. First, assuming that X and Y have a joint Gaussian
distribution. However, we cannot use this method because one of the variables - a set
of class labels - is discrete. Second, modeling distribution of X or Y using parametric
form , for example, mixture of Gaussians [4]. Third, using non-parametric techniques
to estimate distributions of X and Y [5]. The proposed analysis is based on the
second method, where distribution of features is modeled as a Gaussian distribution.
Although it is a strong assumption, we show that results of this analysis are similar
to the results obtained using the third method [5].
The paper is organized as follows. Section 2 describes the experimental setup.
Section 3 describes MAN OVA and presents results of MAN OVA. Section 4 proposes
information theoretic approach for analysis of information in speech and presents
the results. Section 5 compares these results with results from the previous study.
Section 6 describes the summary and conclusions from this work.
2
Experimental Setup
In the previous work [1 , 2], we have analyzed variability in the features using three
databases - HTIMIT, OGI Stories and TIMIT. In this work, we present results
of MAN OVA using OGI Stories database; mainly for the comparison with Yang's
results [5, 6]. English part of OGI Stories database consists of 207 speakers, speaking
for approximately 1 minute each. Each utterance is transcribed at phone level.
Therefore, phone is considered as a source of variability or source of information.
The utterances are not labeled separately by speakers and channels, so we cannot
measure speaker and channel as separate sources. Instead, we assume that different
speakers have used different channels and consider speaker+channel as a single
source of variability or a single source of information.
Figure 1 shows a commonly used time-frequency representation of energy in speech
signal. The y-axis represents frequency, x-axis represents time, and the darkness of
each element shows the energy at a given frequency and time. A spectral vector is
defined by the number of points on the y-axis, S(w, t m ). In this work, this vector
contains 15 points on Bark spectrum. The vector is estimated at every 10 ms using a
25 ms speech segment. It is labeled by the phone and the speaker and channel label
of the corresponding speech segment. A temporal vector is defined by a sequence
of points along time at a given frequency, S(wn, t). In this work, it consists of
50 points each in the past and the future with respect to the current observation
and the observation itself. As the spectral vectors are computed every 10 ms, the
temporal vector represents 1 sec of temporal information. The temporal vectors
are labeled by the phone and the speaker and channel label of the current speech
segment. In this work, the analysis is performed independently using spectral and
temporal vectors.
3
MANOVA
Multivariate analysis of variance (MANOVA) [7] is used to measure the variation
in the data, {X E R n }, with respect to two or more factors. In this work, we use
two factors - phone and speaker+channel. The underline model of MAN OVA is
(1)
where, i = 1"" ,p, represents phones, j = 1"" Be, represents speakers and channels. This equation shows that any feature vector, X ijk , can be approximated using
a sum of X.. , the mean of the data; Xi ., mean of the phone i; Xij., mean of the
speaker and channel j, and phone i; and Eij k, an error in this approximation. Using
~r- l0ms
Ii I
JJ I 1M
I
LL
I
I~I 11 IH~ifl~" III
I'I
~i
I
I I1I11
1
' 1!
I~
I ~ [I
1111
1;1, I
?
?
i
I III
1II1
IHII
I
n
I[l l~
'I
~
11111
I
Il:
I II ,UI
I
~IJ' n'I ~
il [ItJ:ii'
l!I I I'""I~ "'I 'I ~'I
u
1'1
ilJU JI II,I "i...i
"
uJlJ
111 1'111
1111
[III
r1111
III
II I
Temporal Vector
(Temporal Domain)
1111
---:JIIIII_ _ __
Spectral Vector
(Spectral Domain)
Figure 1: Time-frequency representation of logarithmic energies from speech signal
this model, the total covariance can be decomposed as follows
~total = ~p
+ ~s c + ~re sidual
(2)
where
N (X... _
"""' N i
~
~sc
LL -Nij
N (X, ZJ
i
~r esidual
X .. )t (X... - X .. )
-
t
-
-
X .. ) (X,
ZJ - X z. )
j
1""",,,"",,,"",
-
t
-
N ~ ~ ~(Xijk - Xij) (Xijk - Xij)
i
j
k
and, N is the data size and Nijk refers to the number of samples associated with
the particular combination of factors (indicated by the subscript).
The covariance terms are computed as follows. First, all the feature vectors (X)
belonging to each phone i are collected and their mean (Xi) is computed. The
covariance of these phone means, ~p, is the estimate of phone variability. Next, the
data for each speaker and channel j within each phone i is collected and the mean
of the data (X ij ) is computed. The covariance of the means of different speakers
averaged over all phones, ~s c, is the estimate of speaker variability. All the variability in the data is not explained using these sources. The unaccounted sources,
such as context and coarticulation, cause variability in the data collected from
one speaker speaking one phone through one channel. The covariance within each
phone, speaker, and channel is averaged over all phones, speakers, and channels,
and the resulting covariance, ~r e sidual, is the estimate of residual variability.
3.1
Results
Results of MAN OVA are interpreted at two levels - feature element and feature
vector. Results for each feature element are shown in Figure 2. Table 1 shows the
results using the complete feature vector. The contribution of different sources is
calculated as trace (~source )ltrace(~total). Note that this measure cannot be used
to compare variabilities across feature-sets with different number of features. Therefore, we cannot directly compare contribution of variabilities in time and frequency
domains. For comparison, contribution of sources in temporal domain is calculated
Table 1: Contribution of sources in spectral and temporal domains
o contribution
source
pectral Domain Temporal Domain
phone
35.3
4.0
speaker+channel
41.1
30.3
as trace(EtI',source E) /trace(EtI',totaIE) , where
eigenvectors of I',total.
ElOl x 15
is a matrix of 15 leading
In spectral domain, the highest phone variability is between 4-6 Barks. The highest
speaker and channel variability is between 1-2 Barks where phone variability is
the lowest. In temporal domain, phone variability spreads for approximately 250
ms around the current phone. Speaker and channel variability is almost constant
except around the current frame. This deviation is explained by the difference in
t he phonetic context among the phone instances across different speakers. Thus,
features for speakers within a phone differ not only because of different speaker
characteristics but also different phonetic contexts. This deviation is also seen in
the speaker and channel information in the proposed analysis. In the overall results
for each domain, spectral domain has higher variability due to different phones than
temporal domain. It also has higher speaker and channel variability than temporal
domain.
The disadvantage of this analysis is that it is difficult to interpret the results. For
example, how much phone variability is needed for perfect phone recognition? and
is 4% of phone variability in temporal domain significant? In order to answer these
questions, we propose an information theoretic analysis.
Phone
Speaker+Channel
7r,----------------------~
6
7 ,---~----~--------~
,
~
6
5
5 '
Q)
"'
g 4 ,.. ,. .. , _ ,_ , _ ,
..
<1l
_ , _ ,_ ' - ,- '
...
;
'?3
>
2
O L-----~----~------~
5
10
Frequency
(Critical Band Index)
15
-250
0
Time (ms)
Figure 2: Results of analysis of variability
250
4
Information-theoretic Analysis
Results of MAN OVA can not be directly converted to MI because the determinant
of source and residual covariances do not add to the determinant of total covariance.
Therefore, we propose a different formulation for the information theoretic analysis
as follows. Let {X E Rn} be a set of feature vectors, with probability distribution
p(X). Let h(X) be the entropy of X. Let Y = {Y1 , ... , Ym} be a set of different
factors and each Yi be a set of classes within each factor. For example, we can
assume that Y1 = {yf} represents phone factor and each yf represent a phone class.
Lets assume that X has two parts; one completely characterized by Y and another
part, Z , characterized by N(X) ""' N(O, Jn xn ), where J is the identity matrix. Let
J (X; Y) be the MI between X and Y. Assuming that we consider all the possible
factors for our analysis,
J(X;Y) = J(X;Y1, ... , Ym) = h(X)-h(X/Yl , ... ,Ym) = h(X)-h(Z) = D(PIIN) ,
where D() is the kullback-liebler distance [3] between distributions P and N. Using
the chain-rule, the left hand side can be expanded as follows ,
m
J(X; Y1 ,?.?, Yn ) = J(X; Yd
+ J(X; Y2 /Yd + l: J(X; Yi/Yi - l"'"
Y2 , Yd?
(3)
i=3
If we assume that there are only two factors Y1 and Y2 used for the analysis, then
this equation is similar to the decomposition performed using MAN OVA (Equation
2). The term on the left hand side is entropy of X which is the total information
in X that can be explained using Y . This is similar to the left-hand side term in
MANOVA that describes the total variability. On the right hand side, first term
is similar to the phone variability, second term is similar to the speaker variability,
and the last term which calculates the effect of unaccounted factors (Y3 , ... , Y m ) is
similar to the residual variability.
First and second terms on the right hand side of Equation 3 are computed as follows.
Yd = h(X) -
(4)
J(X; Y2 /Yd = h(X/Yd - h(X/Y1, Y2 ).
(5)
h () terms are estimated using parametric approximation to the total and conditional distribution It is assumed that the total distribution of features is a Gaussian
distribution with covariance ~. Therefore, h (X) = ~ log (2net I~I. Similarly, we
assume that the distribution of features of different phones (i) is a Gaussian distribution with covariances ~i' Therefore,
J(X;
h(X/Y1) =
~
h(X/Yd
l: p (y~)Iog (2net I~il
(6)
yiCYi
Finally, we assume that the distribution of features of different phones spoken by
different speakers is also a Gaussian distribution with covariances ~ij. Therefore,
h(X/Y1,Y2 )
=~
p(yLYOlog(2netl~ijl
l:
(7)
y;CY1,y~CY2
Substituting equations 6 and 7 in equations 4 and 5, we get
J(X ' Y;)=~lo
,
1
2
1
J(X;Y2 /Yd = -log
2
g
I~I
IT.Yi CY;
IT?
IT
.
(8)
1~ ,' IP(Yil
I~'IP(Y;)
YicY;'
i
j
Yl CY1 'Y2 CY2
.
j
I~i IP(Yi ' Y2)
(9)
Phone
Speaker +Channel
0.6 ,----~--~-~-------,
0.5
Ul
1.5
ca
-:2:c
,i
,
0.5
\
:2: 0.2 , .. ,_ , .. ,_ ,' ,; -
\
\
-' , - ,- ,- ,- ,_ ,_ ,_ 1- '"
.... ,. - .'..... , -
, _.,
"
0.1
5
10
Frequency
(Critical Band Index)
15
-250
0
250
Time (ms)
Figure 3: Results of information-theoretic analysis
Table 2: Mutual information between features and phone and speaker and channel
labels in spectral and temporal domains
source
phone
speaker+ channel
4 .1
Results
Figure 3 shows the results of information-theoretic analysis in spectral and temporal domain. These results are computed independently for each feature element.
In spectral domain, phone information is highest between 3-6 Barks. Speaker and
channel information is lowest in that range and highest between 1-2 Barks. Since
OGI Stories database was collected over different telephones, speaker+ channel information below 2 Barks ( :=::: 200 Hz ) is due to different telephone channels. In
temporal domain, the highest phone information is at the center (0 ms). It spreads
for approximately 200 ms around the center. Speaker and channel information is
almost constant across t ime except near the center.
Note that the nature of speaker and channel variability also deviates from the constant around the current frame. But, at the current frame , phone variability is
higher than speaker and channel variability. The results of analysis of informat ion show that, at the current frame, phone information is lower than speaker and
channel information. This difference is explained by comparing our MI results with
results from Yang et. al. [6] in the next section.
Table 2 shows the results for the complete feature vector. Note that there are some
practical issues in computing determinants in Equation 4 and 5. They are related
to data insufficiency, specifically, in temporal domain where the feature vector is
101 points and there are approximately 60 vectors per speaker per phone. We ob-
serve that without proper conditioning of covariances, the analysis overestimates
MI (l(X ; Yl , Y2 ) > H(Yl , Y2 )). This is addressed using the condition number to
limit the number of eigenvalues used in the calculation of determinants. Our hypothesis is that in presence of insufficient data, only few leading eigen vectors are
properly estimated. We have use condition number of 1000 to estimate determinant
of ~ and ~i, and condition number of 100 to estimate the determinant of ~ij. The
results show that phone information in spectral domain is 1.6 nats. Speaker and
channel information is 0.5 nats. In temporal domain, phone information is about
1.2 nats. Speaker and channel information is 5.9 nats. Comparison of results from
spectral and temporal domains shows that spectral domain has higher phone information than temporal domain. Temporal domain has higher speaker and channel
information than spectral domain.
Using these results, we can answer the questions raised in Section 3. First question
was how much phone variability is needed for perfect phone recognition? The answer to the question is H(Yd, because the maximum value of leX; Yd is H(Yd?
We compute H(Yl ) using phone priors. For this database, we get H(Yl ) = 3.42
nats, that means we need 3.42 nats of information for perfect phone recognition.
Question about significance of phone information in temporal domain is addressed
by comparing it with information-less MI level. The information-less MI is computed as MI between the current phone label and features at 500 ms in the past
or in the future . From our results, we get information-less MI equal to 0.0013 nats
considering feature at 500 ms in the past, and 0.0010 nats considering features at
500 ms in the future l . The phone information in temporal domain is 1.2 bits that
is greater than both the levels. Therefore it is significant.
5
Results in Perspective
In the proposed analysis, we estimated MI assuming Gaussian distribution for the
features. This assumption is validated by comparing our results with the results
from a study by Yang, et. al.,[6], where MI was computed without assuming any
parametric model for the distribution of features. Note that only entropies can
be directly compared for difference in the estimation technique [3]. However, MI
using Gaussian assumption can be equal to, less or more than the actual MI. In
the comparison of our results with Yang's results , we consider only the nature of
information observed in both studies. The difference in actual MI levels across the
two studies is related to the difference in the estimation techniques.
In spectral domain, Yang's study showed higher phone information between 3-8
Barks. The highest phone information was observed at 4 Barks. Higher speaker
and channel information was observed around 1-2 Barks. In temporal domain, their
study showed that phone information spreads for approximately 200 ms around the
current time frame. Comparison of results from this analysis and our analysis shows
that nature of phone information is similar in both studies. Nature of speaker and
channel information in spectral domain is also similar. We could not compare the
speaker and channel information in temporal domain because Yang's study did not
present these results.
In Section 4.1, we observed difference in the nature of speaker and channel variability, and speaker and channel information at Ii =5 Barks. Comparing MI levels
from our study to those from Yang's study, we observe that Yang's results show that
speaker and channel information at 5 Barks is less that the corresponding phone
information. This is consistent with results of analysis of variability, but not with
lInformation-less MI calculated using Yang et. al. is 0.019 bits
the proposed analysis of information. As mentioned before, this difference is due
to difference in the density estimation techniques used for computing MI. In the
future work, we plan to model the densities using more sophisticated techniques,
and improve the estimation of speaker and channel information.
6
Conclusions
We proposed analysis of information in speech using three sources of information
- language (phone), speaker and channel. Information in speech was measured as
MI between the class labels and the set of features extracted from speech signal.
For example, linguistic information was measured using phone labels and the features. We modeled distribution of features using Gaussian distribution. Thus we
related the analysis to previous proposed analysis of variability in speech. We observed similar results for phone variability and phone information. The speaker
and channel variability and speaker and channel information around the current
frame was different. This was shown to be related to the over-estimation of speaker
and channel information using unimodal Gaussian model. Note that the analysis of
information was proposed because its results have more meaningful interpretations
than results of analysis of variability. For addressing the over-estimation, we plan
to use more complex models ,such as mixture of Gaussians, for computing MI in
the future work.
Acknowledgments
Authors thank Prof. Andrew Fraser from Portland State University for numerous
discussions and helpful insights on this topic.
References
[1] S. S. Kajarekar , N. Malayath and H. Hermansky, "Analysis of sources of variability in speech," in Proc. of EUROSPEECH, Budapest, Hungary, 1999.
[2] S. S. Kajarekar, N. Malayath and H. Hermansky, "Analysis of speaker and
channel variability in speech," in Proc. of ASRU, Colorado, 1999.
[3] T. M. Cover and J. A. Thomas, Elements of Information theory, John Wiley &
Sons, Inc., 1991.
[4] J. A. Bilmes, "Maximum Mutual Information Based Reduction Strategies for
Cross-correlation Based Joint Distribution Modelling ," in Proc. of ICASSP,
Seattle, USA, 1998.
[5] H. Hermansky H. Yang, S. van Vuuren, "Relevancy of Time-Frequency Features
for Phonetic Classification Measured by Mutual Information," in ICASSP '99,
Phoenix, Arizona, USA, 1999.
[6] H. H. Yang, S. Sharma, S. van Vuuren and H. Hermansky, "Relevance of TimeFrequency Features for Phonetic and Speaker-Channel Classification," Speech
Communication, Aug. 2000.
[7] R. V. Hogg and E. A. Tannis, Statistical Analysis and Inference, PRANTICE
HALL, fifth edition, 1997.
| 2161 |@word determinant:6 timefrequency:1 underline:1 relevancy:1 covariance:13 decomposition:1 carry:2 reduction:1 contains:1 past:3 current:10 comparing:4 john:1 vuuren:2 along:1 consists:2 decomposed:1 actual:2 considering:2 lowest:2 interpreted:1 proposing:1 spoken:1 temporal:26 berkeley:1 every:2 y3:1 yn:1 overestimate:1 before:1 engineering:2 insufficiency:1 limit:1 subscript:1 approximately:5 yd:11 range:1 averaged:2 practical:1 acknowledgment:1 nijk:1 refers:1 get:3 cannot:4 context:3 darkness:1 center:3 independently:2 rule:1 insight:1 variation:1 colorado:1 hypothesis:1 element:5 approximated:1 recognition:3 database:5 labeled:3 observed:5 electrical:1 cy:1 ifl:1 highest:6 mentioned:1 ui:1 nats:9 segment:3 serve:1 completely:1 icassp:2 joint:2 sc:1 itself:1 ip:3 sequence:1 eigenvalue:1 net:2 propose:3 budapest:1 hungary:1 seattle:1 perfect:3 andrew:1 measured:9 ij:4 school:1 aug:1 strong:1 differ:1 coarticulation:1 around:7 considered:1 hall:1 substituting:1 estimation:6 proc:3 label:9 eti:2 gaussian:10 asp:1 linguistic:3 validated:1 properly:1 portland:1 modelling:1 mainly:1 helpful:1 inference:1 overall:1 among:1 issue:1 classification:2 proposes:1 plan:2 raised:1 mutual:6 equal:2 represents:6 hermansky:5 future:5 few:1 pectral:1 ime:1 message:1 intra:1 mixture:2 analyzed:1 chain:1 re:1 nij:1 instance:1 modeling:1 cover:1 disadvantage:1 deviation:2 addressing:1 eurospeech:1 answer:3 density:2 international:1 yl:6 ym:3 itj:1 transcribed:1 leading:2 converted:1 sec:1 inc:1 cy2:2 performed:3 timit:1 contribution:6 ii1:1 il:3 variance:2 characteristic:1 bilmes:1 liebler:1 energy:3 frequency:9 associated:1 mi:21 organized:1 sophisticated:1 higher:7 formulation:1 correlation:1 hand:5 yf:2 indicated:1 usa:2 effect:3 y2:11 ogi:6 ll:2 speaker:57 m:12 ijl:1 theoretic:7 complete:2 ji:1 yil:1 phoenix:1 unaccounted:2 conditioning:1 extend:1 interpretation:2 he:1 interpret:2 significant:2 hogg:1 similarly:1 language:3 similarity:1 add:1 multivariate:2 showed:2 perspective:1 phone:66 phonetic:4 yi:5 seen:1 greater:1 sharma:1 signal:6 ii:6 unimodal:1 characterized:2 calculation:1 cross:1 divided:1 fraser:1 iog:1 calculates:1 represent:1 ion:1 separately:1 addressed:2 source:22 hz:1 near:1 yang:11 presence:1 iii:4 wn:1 ul:1 speech:23 speaking:2 cause:1 jj:1 eigenvectors:1 band:2 sachin:2 hynek:2 xij:3 zj:2 estimated:4 per:2 discrete:1 sum:1 almost:2 ob:1 informat:1 bit:3 arizona:1 expanded:1 department:1 combination:1 belonging:1 describes:4 across:4 son:1 explained:4 equation:7 needed:2 gaussians:2 observe:2 spectral:17 eigen:1 jn:1 thomas:1 beaverton:1 prof:1 question:5 lex:1 parametric:4 strategy:1 distance:1 separate:1 thank:1 topic:1 collected:4 assuming:4 modeled:3 index:2 insufficient:1 netl:1 setup:2 difficult:1 trace:3 proper:1 observation:2 variability:48 communication:2 frame:6 rn:1 y1:8 namely:1 below:1 critical:2 residual:3 improve:1 numerous:1 axis:3 utterance:2 deviate:1 prior:1 bark:11 consistent:1 story:4 lo:1 cy1:2 summary:1 last:1 english:1 ovum:7 side:5 institute:1 fifth:1 van:2 calculated:3 xn:1 author:1 commonly:1 kullback:1 assumed:1 xi:2 spectrum:1 table:4 channel:49 nature:5 ca:2 xijk:2 complex:1 domain:32 did:1 significance:1 main:1 spread:3 edition:1 wiley:1 third:2 minute:1 ih:1 easier:1 entropy:3 logarithmic:1 manova:6 eij:1 extracted:3 conditional:1 identity:1 asru:1 man:7 telephone:2 except:2 specifically:1 total:11 experimental:2 ijk:1 meaningful:2 relevance:1 ohsu:1 |
1,276 | 2,162 | Real-Time Monitoring of Complex Industrial
Processes with Particle Filters
Rub?en Morales-Men?endez
Dept. of Mechatronics and Automation
ITESM campus Monterrey
Monterrey, NL M?exico
[email protected]
Nando de Freitas and David Poole
Dept. of Computer Science
University of British Columbia
Vancouver, BC, V6T 1Z4, Canada
nando,poole @cs.ubc.ca
Abstract
This paper discusses the application of particle filtering algorithms to
fault diagnosis in complex industrial processes. We consider two ubiquitous processes: an industrial dryer and a level tank. For these applications, we compared three particle filtering variants: standard particle filtering, Rao-Blackwellised particle filtering and a version of RaoBlackwellised particle filtering that does one-step look-ahead to select
good sampling regions. We show that the overhead of the extra processing per particle of the more sophisticated methods is more than compensated by the decrease in error and variance.
1 Introduction
Real-time monitoring is important in many areas such as robot navigation or diagnosis
of complex systems [1, 2]. This paper considers online monitoring of complex industrial
processes. The processes have a number of discrete states, corresponding to different combinations of faults or regions of qualitatively different dynamics. The dynamics can be very
different based on the discrete states. Even if there are very few discrete states, exact monitoring is computationally unfeasible as the state of the system depends on the history of the
discrete states. However there is a need to monitor these systems in real time to determine
what faults could have occurred.
This paper investigates the feasibility of using particle filtering (PF) for online monitoring.
It also proposes some powerful variants of PF. These variants involve doing more computation per particle for each time step. We wanted to investigate whether we could do real-time
monitoring and whether the extra cost of the more sophisticated methods was worthwhile
in these real-world domains.
2 Classical approaches to fault diagnosis in dynamic systems
Most existing model-based fault diagnosis methods use a technique called analytical redundancy [3]. Real measurements of a process variable are compared to analytically calculated
Visiting Scholar (2000-2003) at The University of British Columbia.
values. The resulting differences, named residuals, are indicative of faults in the process.
Many of these methods rely on simplifications and heuristics [4, 5, 6, 7]. Here, we propose
a principled probabilistic approach to this problem.
3 Processes monitored
We analyzed two industrial processes: an industrial dryer and a level-tank. In each of these,
we physically inserted a sequence of faults into the system and made appropriate measurements. The nonlinear models that we used in the stochastic simulation were obtained
through open-loop step responses for each discrete state [8]. The parametric identification
procedure was guided by the minimum squares error algorithm [9] and validated with the
?Control Station? software [10]. The discrete-time state space representation was obtained
by a standard procedure in control engineering [8].
3.1 Industrial dryer
An industrial dryer is a thermal process that converts electricity to heat. As shown in
Figure 1, we measure the temperature of the output air-flow.
Figure 1: Industrial dryer.
Normal operation corresponds to low fan speed,
open air-flow grill and clean temperature
sensor (we denote this discrete state
). We induced 3 types of fault: faulty
fan,
faulty grill (the grill is closed), and
faulty fan and grill.
3.2 Level tank
Many continuous industrial processes need to control the amount of accumulated material
using level measurement, such as evaporators, distillation columns or boilers. We worked
with a level-tank system that exhibits the dynamic behaviour of these important processes,
see Figure 2. A by-pass pipe and two manual valves (
and ) where used to induce
typical faulty states.
4 Mathematical model
We adopted the following jump Markov linear Gaussian model:
,
!#"$ &%')( &*+
)./ &012)3 &*+54
-
Figure 2: Level-Tank
where ,
denotes the measurements,
denotes the unknown continuous
4 54
is a known control signal,
denotes the unknown discrete
states, *
states (normal operations and faulty conditions). The noise processes are i.i.d Gaussian:
%'
4 and 01
4 . The parameters 4 " 4 - 4 . 4 ('4 3 4
are
4
identified matrices with . ./
for any . The initial states are
. The important thing to notice is that for each realization of 1 , we have
and
a single linear-Gaussian model. If we knew , we could solve for exactly using the
Kalman filter algorithm.
'&
! #" %$
&
The aim of the analysis is to compute the marginal posterior distribution1 of the discrete states ,
. This distribution can be derived from the posterior distribution
4
,
by standard marginalisation. The posterior density satisfies the following recursion
( '& )&
*
&
'& 4 )& ,
&
*
'&
4 '& 2
,
& 2
*
, 54
*
*
4
4
, ,
2
&
(1)
This recursion involves intractable integrals. One, therefore, has to resort to some form of
numerical approximation scheme.
5 Particle filtering
'+-,/& . 4 )+/,/& . 4 % -+ ,/. 10/, 2
In the PF setting, we use a weighted set of samples (particles)
approximate the posterior with the following point-mass distribution
3
to
4'& 54 '& ,
& 5 0 % 7
+/,-. 68:>@9<;-? =AB >@9<;-? A= ( 4'& 54 '& 4
(
0
,-2
)
&
)
&
4
& denotes the Dirac-delta function.
where 6 8 >9C;/? =A B >9C;/? A= D(
Given E
parti
cles F)+/,/& .
4 '+/,-& . 2
10,/2
at time GIH
, approximately distributed according to
NOTATION: For a generic vector J , we adopt the notation JLKM NFOIP#J1K%Q@JRQ%SS%S%Q@J'NUTWV to denote all
the entries of this vector at time X . For simplicity, we use JN to denote both the random variable and its
realisation.
we express continuous probability distributions using YZP\[LJ N T instead of
]_^ P#J'Na`[LJ'NDConsequently,
T and discrete distributions using YbP#J'NUT instead of ]_^ P#J'N_cdJNWT . If these distributions
admit densities with respect to an underlying measure e (counting or Lebesgue), we denote these
densities by fgP#JNWT . For example, when considering the space hji , we will use the Lebesgue measure,
ekcl[LJ'N , so that YZP\[LJNWTFcmfgP#J'NUTL[LJN .
1
( '+-,/& .
4 '+/,-& . 2
,
& 2
, PF enables us to compute E particles '+/,-& . 4 '+/,-& . 0,-2
approximately distributed according to ( )+/,/& . 4 '+/,-& . ,
& , at time G . Since we cannot sample from
the posterior directly, the PF update is accomplished by introducing an appropriate importance proposal distribution ( )& 54 '& from which we can obtain samples. The basic
algorithm, Figure (3), consists of two steps: sequential importance sampling and selection
(see [11] for a detailed derivation). This algorithm uses the transition priors as proposal
distributions; 4 ,
4 2
. For the selection step, we
used a state-of-the-art minimum variance resampling algorithm [12].
'& )& &
Sequential importance sampling step
c Q)SCSCS Q , sample from the transition priors
N
Y P N N
K T and N
Y P\[ N N
K Q N
T
and set M N Q M N O N Q N Q M N K Q M N K S
For c Q)SCSCS Q , evaluate and normalize the importance weights
" f %$ N N
Q N
! N#
For
Selection step
&
M N Q
M N% ')( * K with respect to high/low importance
weights ! N to obtain particles & M N Q M N ' ( * .
K
Multiply/Discard particles
G
Figure 3: PF algorithm at time .
6 Improved Rao-Blackwellised particle filtering
*
'& )&
*
&
& ' & * )& &
'& & '&
* ) &
)&
*
,
4
,
, it is
By considering the factorisation 4 ,
,
4
is Gaussian
possible to design more efficient algorithms. The density
5 ,
& .
and can be computed analytically if we know the marginal posterior density
This density satisfies the alternative recursion
*
* )& 5 ,
& * )&
,
&
*
&
)& *
&
, ,
14
&
, ,
2
*
&
(2)
If equation (1) does not admit a closed-form expression, then equation (2) does not admit one either and sampling-based methods are still required. (Also note that the term
, ,
14
& in equation (2) does not simplify to , & because there is a dependency on past values through .) Now assuming that we can use a weighted set of
samples
4 %
to represent the marginal posterior distribution
&
)&
*
'&
)+/,/& . /+ ,-. 0-, 2
the marginal density of
* 3 4'& ,
&
0
)&
3
0
'& ,
&
50
,-2
%
+/,/. 6 >@9<;-? A= '& &4
is a Gaussian mixture
,+ * 4 '& '& 4 ,
& &@( '& ,
& 5 0
,-2
%
+/,-. * 4 '& ,
& 4 )+/,/& .
that can be computed efficiently with a stochastic bank of Kalman filters. That is, we use
PF to estimate the distribution of and exact computations (Kalman filter) to estimate the
mean and variance of . In particular, we sample and then propagate the mean
and covariance of with a Kalman filter:
$ +/,/.
e N
N
N N
+/,-.
K c
c
c
c
c
"
,
,
0
0
2
, !
where
&
$
P N
TDe N
K P N
T N
P N T N K P N T
P N T P N T
P N T N N K
P N T
P N T P N T
P N TDe N N K P N T N
e N
N K N
N K
P N
T
N K
P $ N $ N
N K T
N N K N N K
P N T
N K
P N T N N K Q
,
& , ,
, ,
&
, $
& 2
, "
,
& & and
,
,
&
0
.
c
N
K
$ N
N K
e N
N
" +/,-.
This is the basis of the RBPF algorithm that was adopted in [13]. Here, we introduce an
extra improvement. Let us expand the expression for the importance weights:
%
* ' & 5 ,
& & * '& 2
,
& * )& 2
4 ,
&
*
(3)
'& ,
&
) &
,
&
5 )& 2
4 ,
&
* , ,
& 2
4 '& * '&
4 ,
& 2
(4)
)&
4 ,
&
)& ,
& '& 2
4 ,
& * '& 2
,
& 2
, states that we are
The proposal choice,
not sampling past trajectories. Sampling past trajectories requires solving an intractable
integral [14].
'&
&
transition prior as proposal distribution: 2
4 ,
. Then, according to equation (4), the importance
weights simplify to the predictive density
the
*We could
)&
14 use
,
&
5
*
*
&
'& &
"! , $# ,
4 %
(5)
However, we can do better by noticing that according to equation (3), the optimal proposal distribution corresponds to the choice 1 )& 2
4 ,
& * )&
4 ,
& . This
%'
, 5 ,
2
4
distribution satisfies Bayes rule:
*
'& 2
4 ,
&
*
&
'& *
&
'&
&
, ,
2
4
4 ,
2
, ,
2
4
5
*
'&
(6)
and, hence, the importance weights simplify to
%'
*
&
, ,
2
4
'&
5
5 '& *
A2
&
, ,
2
4
'&
4
*
(7)
When the number of discrete states is small, say 10 or 100, we can compute the distributions
in equations (6) and (7) analytically. In addition to Rao-Blackwellisation, this leads to
substantial improvements over standard particle filters. Yet, a further improvement can still
be attained.
* '&
& * '&
&
Even when using the optimal importance distribution, there is a discrepancy arising from
the ratio
,
)( 2
,
2
in equation (3). This discrepancy is what causes
the well known problem of sample impoverishment in all particle filters [11, 15]. To circumvent it to a significant extent, we note that the importance weights do not depend on
(we are marginalising over this variable). It is therefore possible to select particles
be
fore the sampling step. That is, one chooses the fittest particles at time
using the
information at time . This observation leads to an efficient algorithm (look-ahead RBPF),
whose pseudocode is shown in Figure 4. Note that for standard PF, Figure 3, the importance weights depend on the sample , thus not permitting selection before sampling.
Selecting particles before sampling results in a richer sample set at the end of each time
step.
G H
G
Kalman prediction step
+/,-.
N_c Q%S%SS)Q compute
e N
N K P N T Q N
N K P N T Q $ N
N K P N T Q N
P N T
For i=1, . . . , N, and for
For i=1 , . . . , N , evaluate and normalize the importance weights
! N
mc f P $ N $ KM N K Q
M N K TFc i P $ N
N K P N T Q N
P N TT-f P N N
K T
A* K
Selection step
Multiply/Discard particles & e N
K Q N
K Q
M N K ' ( * K with respect to high/low
importance weights ! N to obtain particles & e N K Q N K Q M N K ' ( * .
K
Sequential importance sampling step
Kalman prediction. For i=1, . . . , N, and for N_c QS%S)S%Q using the resampled
information, re-compute
e N
N K P NWT Q N
N K P NUT Q $ N
N K P NWT Q N
P NUT
For N c Q)S%SS%Q compute
f P N
M N K Q $ KM NUT " P $ N
N K P NWT Q N
P NDTT-f P N N
K T
Sampling step
N
f P N
M N K Q $ KM NUT
Updating step
For i=1 , . . . , N, use one step of the Kalman recursion
the suffi
to
compute
cient statistics & e N Q N ' given & e N N K P N T Q N N K P N T ' .
&
G
GH
Figure 4: look-ahead RBPF algorithm at time . The
algorithm uses an optimal proposal
distribution. It also selects particles from time
using the information at time .
G
7 Results
The results are shown in Figures 5 and 6. The left graphs show error detection versus
computing time per time-step (the signal sampling time was 1 second). The right graphs
show the error detection versus number of particles. The error detection represents how
many discrete states were not identified properly, and was calculated for 25 independent
runs (1,000 time steps each). The graphs show that look-ahead RBPF works significantly
better (low error rate and very low variance). This is essential for real-time operation with
low error rates.
100
100
90
90
Real time
80
80
% Error Detection
% Error Detection
70
60
50
40
RBPF
30
PF
60
50
RBPF
PF
40
30
20
10
70
20
la?RBPF
10
la?RBPF
0 ?3
10
?2
10
?1
10
0
10
Computing time per timestep (=) sec
1
10
0
0
10
1
2
10
10
Number of particles
3
10
Figure 5: Industrial dryer: error detection vs computing time and number of particles.
The graphs also show that even for 1 particle, look-ahead RBPF is able to track the discrete
state. The reason for this is that the sensors are very accurate (variance=0.01). Consequently, the distributions are very peaked and we are simply tracking the mode. Note that
look-ahead RBPF is the only filter that uses the most recent information in the proposal
distribution. Since the measurements are very accurate, it finds the mode easily. We repeated the level-tank experiments with noisier sensors (variance=0.08) and obtained the
results shown in Figure 7. Noisier sensors, as expected, reduce the accuracy of look-ahead
RBPF with a small number of particles. However, it is still possible to achieve low error
rates in real-time. Since modern industrial and robotic sensors tend to be very accurate, we
conclude that look-ahead RBPF has great potential.
Acknowledgments
Ruben Morales-Men?endez was partly supported by the Government of Canada (ICCS) and
UBC CS department. David Poole and Nando de Freitas are supported by NSERC
References
[1] J Chen and J Howell. A self-validating control system based approach to plant fault detection
and diagnosis. Computers and Chemical Engineering, 25:337?358, 2001.
[2] S Thrun, J Langford, and V Verma. Risk sensitive particle filters. In S Becker T. G Dietterich and Z Ghahramani, editors, Advances in Neural Information Processing Systems 14,
Cambridge, MA, 2002. MIT Press.
[3] J Gertler. Fault detection and diagnosis in engineering systems. Marcel Dekker, Inc., 1998.
[4] P Frank and X Ding. Survey of robust residual generation and evaluation methods in observerbased fault detection systems. J. Proc. Cont, 7(6):403?424, 1997.
[5] P Frank, E Alcorta Garcia, and B.Koppen-Seliger. Modelling for fault detection and isolation
versus modelling for control. Mathematics and computers in simulation, 53:259?271, 2000.
[6] P Frank. Fault diagnosis in dynamic systems using analytical and knowledge-based redundancy
? a survey and some new results. Automatica, 26:459?474, 1990.
[7] R Isermann. Supervision, fault-detection and fault-diagnosis methods - an introduction. Control
engineering practice, 5(5):639?652, 1997.
100
100
90
90
80
PF
Real time
% Error Detection
% Error Detection
80
70
60
50
RBPF
40
RBPF
60
PF
50
40
30
30
20
20
10
70
la?RBPF
0 ?3
10
la?RBPF
10
?2
?1
10
0
10
0
1
10
0
1
10
10
2
10
10
Number of particles
Computing time per timestep (=) sec
Figure 6: Level-tank (accurate sensors): error detection vs computing time and number of
particles.
100
100
90
90
Real time
80
80
PF
70
% Error Detection
% Error Detection
70
60
RBPF
50
40
30
20
60
40
30
20
la?RBPF
la?RBPF
10
PF
RBPF
50
10
0
0
?3
10
?2
10
?1
10
0
10
Computing time per timestep (=) sec
1
10
0
10
1
2
10
10
3
10
Number of particles
Figure 7: Level-tank (noisy sensors): error detection vs computing time and number of
particles.
[8] K Ogata. Discrete-Time Control Systems. Prentice Hall, second edition, 1995.
[9] L Ljung. System Identification: Theory for the User. Prentice-Hall, 1987.
[10] D Cooper. Control Station. University of Connecticut, third edition, 2001.
[11] A Doucet, N de Freitas, and N J Gordon, editors. Sequential Monte Carlo Methods in Practice.
Springer-Verlag, 2001.
[12] G Kitagawa. Monte Carlo filter and smoother for non-Gaussian nonlinear state space models.
Journal of Computational and Graphical Statistics, 5:1?25, 1996.
[13] N de Freitas. Rao-Blackwellised particle filtering for fault diagnosis. In IEEE aerospace conference, 2001.
[14] C Andrieu, A Doucet, and E Punskaya. Sequential Monte Carlo methods for optimal filtering. In
A Doucet, N de Freitas, and N J Gordon, editors, Sequential Monte Carlo Methods in Practice.
Springer-Verlag, 2001.
[15] M Pitt and N Shephard. Filtering via simulation: auxiliary particle filters. Journal of the
American statistical association, 94(446):590?599, 1999.
| 2162 |@word version:1 dekker:1 open:2 simulation:3 propagate:1 covariance:1 gertler:1 initial:1 selecting:1 bc:1 past:3 freitas:5 existing:1 yet:1 numerical:1 enables:1 wanted:1 update:1 resampling:1 v:3 indicative:1 mathematical:1 consists:1 overhead:1 introduce:1 expected:1 valve:1 pf:14 considering:2 campus:1 notation:2 underlying:1 mass:1 what:2 blackwellised:3 exactly:1 connecticut:1 control:9 before:2 engineering:4 approximately:2 acknowledgment:1 practice:3 procedure:2 area:1 significantly:1 induce:1 unfeasible:1 cannot:1 selection:5 faulty:5 risk:1 prentice:2 compensated:1 survey:2 simplicity:1 factorisation:1 rule:1 parti:1 user:1 exact:2 us:3 updating:1 v6t:1 inserted:1 ding:1 region:2 decrease:1 principled:1 substantial:1 dynamic:5 depend:2 solving:1 predictive:1 basis:1 easily:1 derivation:1 boiler:1 heat:1 monte:4 cles:1 whose:1 heuristic:1 richer:1 solve:1 say:1 statistic:2 noisy:1 online:2 sequence:1 analytical:2 propose:1 loop:1 realization:1 achieve:1 j1k:1 fittest:1 dirac:1 normalize:2 shephard:1 auxiliary:1 c:2 involves:1 marcel:1 guided:1 filter:11 stochastic:2 nando:3 material:1 government:1 behaviour:1 scholar:1 kitagawa:1 hall:2 normal:2 great:1 pitt:1 adopt:1 a2:1 proc:1 suffi:1 ndt:1 sensitive:1 weighted:2 mit:1 raoblackwellised:1 sensor:7 gaussian:6 aim:1 validated:1 derived:1 improvement:3 properly:1 modelling:2 industrial:12 accumulated:1 lj:5 expand:1 selects:1 tank:8 impoverishment:1 proposes:1 art:1 marginal:4 sampling:12 represents:1 look:8 peaked:1 discrepancy:2 simplify:3 realisation:1 few:1 gordon:2 modern:1 lebesgue:2 detection:17 investigate:1 multiply:2 evaluation:1 navigation:1 analyzed:1 nl:1 mixture:1 accurate:4 integral:2 re:1 column:1 rao:4 electricity:1 cost:1 introducing:1 entry:1 dependency:1 chooses:1 density:8 probabilistic:1 na:1 admit:3 resort:1 american:1 potential:1 de:5 sec:3 automation:1 inc:1 depends:1 closed:2 doing:1 bayes:1 square:1 air:2 accuracy:1 variance:6 efficiently:1 identification:2 mc:1 carlo:4 monitoring:6 trajectory:2 fore:1 howell:1 history:1 manual:1 monitored:1 knowledge:1 ubiquitous:1 sophisticated:2 attained:1 response:1 improved:1 marginalising:1 langford:1 nonlinear:2 mode:2 dietterich:1 andrieu:1 analytically:3 hence:1 chemical:1 nut:5 n_c:2 self:1 temperature:2 pseudocode:1 association:1 occurred:1 measurement:5 distillation:1 significant:1 cambridge:1 mathematics:1 z4:1 particle:35 robot:1 supervision:1 posterior:7 recent:1 discard:2 verlag:2 fault:16 accomplished:1 minimum:2 determine:1 signal:2 smoother:1 dept:2 permitting:1 feasibility:1 prediction:2 variant:3 basic:1 physically:1 represent:1 proposal:7 addition:1 extra:3 marginalisation:1 induced:1 tend:1 validating:1 thing:1 flow:2 counting:1 isolation:1 identified:2 reduce:1 grill:4 whether:2 expression:2 rmm:1 becker:1 cause:1 detailed:1 involve:1 amount:1 gih:1 notice:1 delta:1 arising:1 per:6 fgp:1 hji:1 track:1 diagnosis:9 discrete:14 express:1 redundancy:2 monitor:1 clean:1 timestep:3 graph:4 convert:1 run:1 noticing:1 powerful:1 nwt:5 named:1 rub:1 investigates:1 resampled:1 simplification:1 fan:3 ahead:8 worked:1 software:1 speed:1 department:1 according:4 combination:1 dryer:6 jlk:1 computationally:1 equation:7 discus:1 know:1 end:1 adopted:2 operation:3 worthwhile:1 generic:1 appropriate:2 endez:2 alternative:1 denotes:4 graphical:1 ghahramani:1 classical:1 parametric:1 visiting:1 exhibit:1 mx:1 thrun:1 considers:1 extent:1 reason:1 assuming:1 kalman:7 cont:1 ratio:1 frank:3 design:1 unknown:2 observation:1 markov:1 thermal:1 station:2 canada:2 david:2 required:1 pipe:1 aerospace:1 able:1 poole:3 distribution1:1 rely:1 circumvent:1 residual:2 recursion:4 scheme:1 columbia:2 prior:3 ruben:1 icc:1 vancouver:1 plant:1 ljung:1 men:2 generation:1 filtering:11 versus:3 rbpf:20 editor:3 bank:1 verma:1 morale:2 supported:2 blackwellisation:1 distributed:2 calculated:2 world:1 transition:3 qualitatively:1 made:1 jump:1 approximate:1 doucet:3 robotic:1 automatica:1 conclude:1 knew:1 continuous:3 robust:1 ca:1 complex:4 domain:1 noise:1 edition:2 repeated:1 cient:1 en:1 cooper:1 third:1 ogata:1 british:2 intractable:2 essential:1 sequential:6 importance:14 chen:1 garcia:1 simply:1 nserc:1 tracking:1 springer:2 ubc:2 corresponds:2 satisfies:3 ma:1 consequently:1 typical:1 called:1 pas:1 partly:1 la:6 select:2 noisier:2 evaluate:2 |
1,277 | 2,163 | One-Class LP Classifier for Dissimilarity
Representations
1
El?zbieta P?ekalska1 , David M.J.Tax2 and Robert P.W. Duin1
Delft University of Technology, Lorentzweg 1, 2628 CJ Delft, The Netherlands
2
Fraunhofer Institute FIRST.IDA, Kekul?str.7, D-12489 Berlin, Germany
[email protected],[email protected]
Abstract
Problems in which abnormal or novel situations should be detected can
be approached by describing the domain of the class of typical examples. These applications come from the areas of machine diagnostics,
fault detection, illness identification or, in principle, refer to any problem where little knowledge is available outside the typical class. In this
paper we explain why proximities are natural representations for domain
descriptors and we propose a simple one-class classifier for dissimilarity
representations. By the use of linear programming an efficient one-class
description can be found, based on a small number of prototype objects.
This classifier can be made (1) more robust by transforming the dissimilarities and (2) cheaper to compute by using a reduced representation set.
Finally, a comparison to a comparable one-class classifier by Campbell
and Bennett is given.
1 Introduction
The problem of describing a class or a domain has recently gained a lot of attention, since it
can be identified in many applications. The area of interest covers all the problems, where
the specified targets have to be recognized and the anomalies or outlier instances have to
be detected. Those might be examples of any type of fault detection, abnormal behavior,
rare illnesses, etc. One possible approach to class description problems is to construct oneclass classifiers (OCCs) [13]. Such classifiers are concept descriptors, i.e. they refer to all
possible knowledge that one has about the class.
An efficient OCC built in a feature space can be found by determining a minimal volume
hypersphere around the data [14, 13] or by determining a hyperplane such that it separates
the data from the origin as well as possible [11, 12]. By the use of kernels [15] the data is
implicitly mapped into a higher-dimensional inner product space and, as a result, an OCC in
the original space can yield a nonlinear and non-spherical boundary; see e.g. [15, 11, 12, 14].
Those approaches are convenient for data already represented in a feature space. In some
cases, there is, however, a lack of good or suitable features due to the difficulty of defining
them, as e.g. in case of strings, graphs or shapes. To avoid the definition of an explicit
feature space, we have already proposed to address kernels as general proximity measures
[10] and not only as symmetric, (conditionally) positive definite functions of two variables
[2]. Such a proximity should directly arise from an application; see e.g. [8, 7]. Therefore,
our reasoning starts not from a feature space, like in case of the other methods [15, 11, 12,
14], but from a given proximity representation. Here, we address general dissimilarities.
The basic assumption that an instance belongs to a class is that it is similar to examples
within this class. The identification procedure is realized by a proximity function equipped
with a threshold, determining whether an instance is a class member or not. This proximity
function can be e.g. a distance to an average representative, or a set of selected prototypes. The data represented by proximities is thus more natural for building the concept
descriptors, i.e. OCCs, since the proximity function can be directly built on them.
In this paper, we propose a simple and efficient OCC for general dissimilarity representations, discussed in Section 2, found by the use of linear programming (LP). Section 3
presents our method together with a dissimilarity transformation to make it more robust
against objects with large dissimilarities. Section 4 describes the experiments conducted,
and discusses the results. Conclusions are summarized in Section 5.
2 Dissimilarity representations
Although a dissimilarity measure D provides a flexible way to represent the data, there
are some constraints. Reflectivity and positivity conditions are essential to define a proper
measure; see also [10]. For our convenience, we also adopt the symmetry requirement.
We do not require that D is a strict metric, since non-metric dissimilarities may naturally
be found when shapes or objects in images are compared e.g. in computer vision [4, 7].
Let z and pi refer to objects to be compared. A dissimilarity representation can now be
seen as a dissimilarity kernel based on the representation set R = {p 1 , .., pN } and realized
by a mapping D(z, R) : F ? RN , defined as D(z, R) = [D(z, p1 ) . . . D(z, pN )]T . R
controls the dimensionality of a dissimilarity space D(?, R). Note also that F expresses a
conceptual space of objects, not necessarily a feature space. Therefore, to emphasize that
objects, like z or pi , might not be feature vectors, they will not be printed in bold.
The compactness hypothesis (CH) [5] is the basis for object recognition. It states that
similar objects are close in their representations. For a dissimilarity measure D, this means
that D(r, s) is small if objects r and s are similar.If we demand that D(r, s) = 0, if and only
if the objects r and s are identical, this implies that they belong to the same class. This can
be extended by assuming that all objects s such that D(r, s) < ?, for a sufficient small ?, are
so similar to r that they are members of the same class. Consequently, D(r, t) ? D(s, t) for
other objects t. Therefore, for dissimilarity representations satisfying the above continuity,
the reverse of the CH holds: objects similar in their representations are similar in reality
and belong, thereby, to the same class [6, 10].
Objects with large distances are assumed to be dissimilar. When the set R contains objects
from the class of interest, then objects z with large D(z, R) are outliers and should be
remote from the origin in this dissimilarity space. This characteristic will be used in our
OCC. If the dissimilarity measure D is a metric, then all vectors D(z, R), lie in an open
prism (unbounded from above1), bounded from below by a hyperplane on which the objects
from R are. In principle, z may be placed anywhere in the dissimilarity space D(?, R) only
if the triangle inequality is completely violated. This is, however, not possible from the
practical point of view, because then both the CH and its reverse will not be fulfilled.
Consequently, this would mean that D has lost its discriminating properties of being small
for similar objects. Therefore, the measure D, if not a metric, has to be only slightly nonmetric (i.e. the triangle inequalities are only somewhat violated) and, thereby, D(z, R) will
still lie either in the prism or in its close neigbourhood.
1
the prism is bounded if D is bounded
3 The linear programming dissimilarity data description
To describe a class in a non-negative dissimilarity space, one could minimize the volume of
the prism, cut by a hyperplane P : w T D(z, R) = ? that bounds the data from above2 (note
that non-negative dissimilarities impose both ? ? 0 and wi ? 0). However, this might be not
a feasible task. A natural extension is to minimize the volume of a simplex with the main
vertex being the origin and the other vertices v j resulting from the intersection of P and
the axes of the dissimilarity space (v j is a vector of all zero elements except for vji = ?/wi ,
given that wi 6= 0). Assume now that there are M non-zero weights of the hyperplane P , so
effectively, P is constructed in a RM . From geometry we know that the volume V of such
a simplex can be expressed as V = (VBase /M !) ? (?/||w||2 ), where VBase is the volume of
the base, defined by the vertices v j . The minimization of h = ?/||w||2 , i.e. the Euclidean
distance from the origin to P , is then related to the minimization of V .
Let {D(pi , R)}N
i=1 , N = |R| be a dissimilarity representation, bounded by a hyperplane P ,
i.e. wT D(pi , R) ? ? for i = 1, . . . , N , such that the Lq distance to the origin dq (0, P ) =
?/||w||p is the smallest (i.e. q satisfies 1/p + 1/q = 1 for p ? 1) [9]. This means that P can
be determined by minimizing ? ? ||w||p . However, when we require ||w||p = 1 (to avoid
any arbitrary scaling of w), the construction of P can be solved by the minimization of ?
only. The mathematical programming formulation of such a problem is [9, 1]:
min ?
(1)
s.t.
wT D(pi , R) ? ?,
i = 1, 2, .., N,
||w||p = 1, ? ? 0.
If p = 2, then P is found such that h is minimized, yielding a quadratic optimization problem. A much simpler
? LP formulation, realized for p = 1, is of our interest. Knowing that
||w||2 ? ||w||1 ? M||w||2 and by?the assumption of ||w||1 = 1, after simple calculations,
we find that ? ? h = ?/||w||2 ? M ?. Therefore, by minimizing d? (0, P ) = ?, (and
||w||1 = 1), h will be bounded and the volume of the simplex considered, as well.
By the above reasoning and (1), a class represented by dissimilarities can be characterized
by a linear proximity function with the weights w and the threshold ?. Our one-class
classifier CLPDD , Linear Programming Dissimilarity-data Description, is then defined as:
CLPDD (D(z, ?)) = I(
X
wj D(z, pj ) ? ?),
(2)
wj 6=0
where I is the indicator function. The proximity function is found as the solution to a soft
margin formulation (which is a straightforward extension of the hard margin case) with
? ? (0, 1] being the upper bound on the outlier fraction for the target class:
PN
min ? + ? 1N
i=1 ?i
(3)
s.t. wT D(pi , R) ? ? + ?i , i = 1, 2, .., N
P
j wj = 1, wj ? 0, ? ? 0, ?i ? 0.
In the LP formulations, sparse solutions are obtained, meaning that only some w j are positive. Objects corresponding to such non-zero weights, will be called support objects (SO).
The left plot of Fig. 1 is a 2D illustration of the LPDD. The data is represented in a metric
dissimilarity space, and by the triangle inequality the data can only be inside the prism
indicated by the dashed lines. The LPDD boundary is given by the hyperplane, as close to
the origin as possible (by minimizing ?), while still accepting (most) target objects. By the
discussion in Section 2, the outliers should be remote from the origin.
Proposition. In (3), ? ? (0, 1] is the upper bound on the outlier fraction for the target class,
i.e. the fraction of objects that lie outside the boundary; see also [11, 12]. This means that
1 PN
i=1 (1 ? CLPDD (D(pi , ?)) ? ?.
N
2
P is not expected to be parallel to the prism?s bottom hyperplane
D(. ,p j)
LPDD: min?
T
LPSD: min 1/N sum k(w K(xk ,S) + ? )
K(.,x j)
?
1
??
|| w ||1 = 1
w
0
Dissimilarity space
D(.,p i)
w
0
Similarity space
1
K(.,x i)
Figure 1: Illustrations of the LPDD in the dissimilarity space (left) and the LPSD in the similarity
space (right). The dashed lines indicate the boundary of the area which contains the genuine objects.
The LPDD tries to minimize the max-norm distance from the bounding hyperplane to the origin,
while the LPSD tries to attract the hyperplane towards the average of the distribution.
The proof goes analogously to the proofs given in [11, 12]. Intuitively, the proof
P follows
this: assume we have found a solution of (3). If ? is increased slightly, the term i ?i in the
objective function will change proportionally to the number of points that have non-zero ? i
(i.e. the outlier objects). At the optimum of (3) it has to hold that N ? ? #outliers.
Scaling dissimilarities. If D is unbounded, then some atypical objects of the target class
(i.e. with large dissimilarities) might badly influence the solution of (3). Therefore, we
propose a nonlinear, monotonous transformation of the distances to the interval [0, 1] such
that locally the distances are scaled linearly and globally, all large distances become close to
1. A function with such properties is the sigmoid function (the hyperbolical tangent can also
be used), i.e. Sigm(x) = 2/(1 + e?x/s ) ? 1, where s controls the ?slope? of the function,
i.e. the size of the local neighborhoods. Now, the transformation can be applied in an
element-wise way to the dissimilarity representation such that Ds (z, pi ) = Sigm(D(z, pi )).
Unless stated otherwise, the CLPDD will be trained on Ds .
A linear programming OCC on similarities. Recently, Campbell and Bennett have
proposed an LP formulation for novelty detection [3]. They start their reasoning from
a feature space in the spirit of positive definite kernels K(S, S) based on the set
S = {x1 , .., xN }. They restricted themselves to the (modified) RBF kernels, i.e. for
2
2
K(xi , xj ) = e?D(xi ,xj ) /2 s , where D is either Euclidean or L1 (city block) distance.
In principle, we will refer to RBFp , as to the ?Gaussian? kernel based on the Lp distance.
Here, to be consistent with our LPDD method, we rewrite their soft-margin LP formulation (a hard margin formulation is then obvious), to include a trade-off parameter ? (which
lacks, however, the interpretation as given in the LPDD), as follows:
PN
PN
min N1 i=1 (wT K(xi , S) + ?) + ? 1N
i=1 ?i
T
(4)
s.t.
w K(xi , S) + ? ? ??i , i = 1, 2, .., N
P
j wj = 1, wj ? 0, ?i ? 0.
Since K can be any similarity representation, for simplicity, we will call this method Linear
Programming Similarity-data Description (LPSD). The CLPSD is then defined as:
CLPSD (K(z, ?)) = I(
X
wj K(z, xj ) + ? ? 0).
(5)
wj 6=0
In the right plot of Fig. 1, a 2D illustration of the LPSD is shown. Here, the data is represented in a similarity space, such that all objects lie in a hypercube between 0 and 1.
Objects remote from the representation objects will be close to the origin. The hyperplane
is optimized to have minimal average output for the whole target set. This does not necessarily mean a good separation from the origin or a small volume of the OCC, possibly
resulting in an unnecessarily high outlier acceptance rate.
LPDD on the Euclidean representation
s = 0.3
1.5
s = 0.4
1.5
s = 0.5
1.5
s=1
1.5
1
1
1
1
0.5
0.5
0.5
0.5
0.5
0
0
0
0
0
?0.5
?0.5
?0.5
?0.5
?0.5
?0.5
0
0.5
1
?0.5
0
0.5
1
?0.5
0
0.5
1
?0.5
0
0.5
s=3
1.5
1
1
?0.5
0
0.5
1
LPSD based on RBF2
s = 0.3
1.5
s = 0.4
1.5
s = 0.5
1.5
s=1
1.5
1
1
1
1
0.5
0.5
0.5
0.5
0.5
0
0
0
0
0
?0.5
?0.5
?0.5
?0.5
?0.5
?0.5
0
0.5
1
?0.5
0
0.5
1
?0.5
0
0.5
1
?0.5
0
0.5
s=3
1.5
1
1
?0.5
0
0.5
1
Figure 2: One-class hard margin LP classifiers for an artificial 2D data. From left to right, s takes
the values of 0.3d, 0.4d, 0.5d, d, 3d, where d is the average distance. Support objects are marked by
squares.
Extensions. Until now, the LPDD and LPSD were defined for square (dis)similarity matrices. If the computation of (dis)similarities is very costly, one can consider a reduced
representation set Rred ? R, consisting of n << N objects. Then, a dissimilarity or similarity representations are given as rectangular matrices D(R, Rred ) or K(S, Sred ), respectively.
Both formulations (3) and (4) remain the same with the only change that R/S is replaced by
Rred /Sred . An another reason to consider reduced representations is the robustness against
outliers. How to choose such a set is beyond the scope of this paper.
4 Experiments
Artificial datasets. First, we illustrate the LPDD and the LPSD methods on two artificial
datasets, both originally created in a 2D feature space. The first dataset contains two clusters with objects represented by Euclidean distances. The second dataset contains one uniform, square cluster and it is contaminated with three outliers. The
P objects are represented
by a slightly non-metric L0.95 dissimilarity (i.e. d0.95 (x, y) = [ i (xi ?yi )0.95 ]1/0.95 ). In
Fig. 2, the first dataset together with the decision boundaries of the LPDD and the LPSD
in the theoretical input space are shown. The parameter s used in all plots refers either to
the scaling parameter in the sigmoid function for the LPDD (based on D s ) or to the scaling
parameter in the RBF kernel. The pictures show similar behavior of both the LPDD and
the LPSD; the LPDD tends to be just slightly more tight around the target class.
LPDD on the Euclidean representation
1
? = 0.1; s = 0.2; e = 0
1
? = 0.1; s = 0.29; e = 0.04
1
? = 0.1; s = 0.46; e = 0.06
1
? = 0.1; s = 0.87; e = 0.06
1
0.5
0.5
0.5
0.5
0.5
0
0
0
0
0
0
1
0.5
1
0
0.5
1
0
0.5
1
0
LPSD based on RBF2
? = 0.1; s = 0.2; e = 0.04
1
? = 0.1; s = 0.29; e = 0.06
1
? = 0.1; s = 0.46; e = 0.06
1
0.5
1
? = 0.1; s = 0.87; e = 0.06
0
1
0.5
0.5
0.5
0.5
0.5
0
0
0
0
0
0
0.5
1
0
0.5
1
0
0.5
1
0
0.5
1
? = 0.1; s = 2.3; e = 0.08
0.5
1
? = 0.1; s = 2.3; e = 0.08
0
0.5
1
Figure 3: One-class LP classifiers, trained with ? = 0.1 for an artificial uniformly distributed 2D data
with 3 outliers. From left to right s takes the values of 0.7dm , dm , 1.6dm , 3dm , 8dm , where dm is
the median distance of all the distances. e refers to the error on the target set. Support objects are
marked by squares.
LPDD on the L0.95 representation
1
? = 0.1; s = 0.26; e = 0
1
? = 0.1; s = 0.37; e = 0.04
1
? = 0.1; s = 0.59; e = 0.06
1
? = 0.1; s = 1.1; e = 0.08
1
0.5
0.5
0.5
0.5
0.5
0
0
0
0
0
0
1
0.5
1
0
0.5
1
0
0.5
1
0
0.5
1
0
LPSD based on RBF0.95
? = 0.1; s = 0.26; e = 0
1
? = 0.1; s = 0.37; e = 0.04
1
? = 0.1; s = 0.59; e = 0.06
1
? = 0.1; s = 1.1; e = 0.08
1
0.5
0.5
0.5
0.5
0.5
0
0
0
0
0
0
0.5
1
0
0.5
1
0
0.5
1
0
? = 0.1; s = 3; e = 0.08
0.5
0.5
1
? = 0.1; s = 3; e = 0.06
1
0
0.5
1
Figure 4: One-class LP classifiers for an artificial 2D data. The same setting as in Fig.3 is used, only
for the L0.95 non-metric dissimilarities instead of the Euclidean ones. Note that the median distance
has changed, and consequently, the s values, as well.
1
LPDD
1
0.5
0.5
0
0
1
LPSD
? = 0.1; s = 0.41; e = 0.08
0
0.5
1
? = 0.1; s = 0.41; e = 0.06
1
? = 0.1; s = 1; e = 0.08
0
0.5
? = 0.1; s = 1; e = 0.08
1
? = 0.1; s = 0.4; e = 0.06
1
0.5
0.5
0
0
1
1
0
0.5
1
? = 0.1; s = 0.4; e = 0.08
1
0.5
0.5
0.5
0.5
0
0
0
0
0
0.5
1
0
0.5
1
0
0.5
1
? = 0.1; s = 1; e = 0.08
0
0.5
? = 0.1; s = 1; e = 0.08
0
0.5
1
1
Figure 5: One-class LP classifiers, trained with ? = 0.1, for an artificial uniformly distributed 2D
data with 3 outliers, given by the L0.95 non-metric rectangular 50?6 dissimilarity representations.
The upper row shows the LPDD?s results and bottom row shows the LPSD?s results with the kernel
RBF0.95 . The objects of the reduced sets Rred and Sred are marked by triangles. Note that they differ
from left to right. e refers to the error on the target set. Support objects are marked by squares.
This becomes more clear in Fig. 3 and 4, where three outliers lying outside a single uniformly distributed cluster should be ignored when an OCC with a soft margin is trained.
From these figures, we can observe that the LPDD gives a tighter class description, which
is more robust against the scaling parameter and more robust against outliers, as well. The
same is observed when L0.95 dissimilarity is used instead of the Euclidean distances.
Fig. 5 presents some results for the reduced representations, in which just 6 objects are
randomly chosen for the set Rred . In the left four plots, Rred contains objects from the
uniform cluster only, and both methods perform equally well. In the right four plots, on
the other hand, Rred contains an outlier. It can be judged that for a suitable scaling s, no
outliers become support objects in the LPDD, which is often a case for the LPSD; see also
Fig. 4 and 3. Also, a crucial difference between the LPDD and LPSD can be observed w.r.t.
the support objects. In case of the LPSD (applied to a non-reduced representation), they lie
on the boundary, while in case of the LPDD, they tend to be ?inside? the class.
Condition monitoring. Fault detection is an important problem in the machine diagnostics: failure to detect faults can lead to machine damage, while false alarms can lead
to unnecessary expenses. As an example, we will consider a detection of four types of
fault in ball-bearing cages, a dataset [16] considered in [3]. Each data instance consists
of 2048 samples of acceleration taken with a Bruel and Kjaer vibration analyser. After
pre-processing with a discrete Fast Fourier Transform, each signal is characterized by 32
attributes. The dataset consists of five categories: normal behavior (NB), corresponding
Table 1: The errors of the first and second kind (in %) of the LPDD and LPSD on two dissimilarity
representations for the ball-bearing data. The reduced representations are based on 180 objects.
LPDD
LPDD-reduced
LPSD
LPSD-reduced
LPDD
LPDD-reduced
LPSD
LPSD-reduced
Euclidean representation
Method
Optimal s # of SO
NB T1
200.4
10
1.4 0.0
65.3
17
1.1 0.0
320.0
8
1.3 0.0
211.2
6
0.6 0.0
L1 dissimilarity representation
Method
Optimal s # of SO
NB T1
566.3
12
1.3 0.0
329.5
10
1.3 0.0
1019.3
8
0.9 0.0
965.7
5
0.3 0.0
Error
T2
45.0
20.2
46.7
39.9
T3
69.8
47.5
71.7
67.1
T4
70.0
50.9
74.5
69.5
Error
T2
1.6
2.3
2.2
3.5
T3
20.9
18.7
27.9
26.3
T4
19.8
16.9
27.2
27.5
to measurements made from new ball-bearings, and four types of anomalies, say, T 1 ? T4 ,
corresponding either to the damaged outer race or cages or a badly worn ball-bearing. To
compare our LPDD method with the LPSD method, we performed experiments in the same
way, as described in [3], making use of the same training set, and independent validation
and test sets; see Fig. 6.
The optimal values of s were found for both LPDD and
Train Valid. Test
LPSD methods by the use of the validation set on the EuNB
913
913
913
clidean and L1 dissimilarity representations. The results
T
747
747
1
are presented in Table 1. It can be concluded that the
T2
913
996
L1 representation is far more convenient for the fault deT3
996
tection, especially if we look at the fault type T3 and T4
T4
996
which were unseen in the validation process. The LPSD
performs better on normal instances (yields a smaller erFigure 6: Fault detection data.
ror) than the LPDD. This is to be expected, since the
boundary is less tight, by which less support objects (SO) are needed. On the contrary, the
LPSD method deteriorates w.r.t. the outlier detection. Note also that the reduced representation, based on randomly chosen 180 target objects (? 20%) mostly yields significantly
better performances in outlier detection for the LPDD, and in target acceptance for the
LPSD. Therefore, we can conclude that if a failure in the fault detection has higher costs
than the cost of misclassifying target objects, then our approach should be recommended.
5 Conclusions
We have proposed the Linear Programming Dissimilarity-data Description (LPDD) classifier, directly built on dissimilarity representations. This method is efficient, which means
that only some objects are needed for the computation of dissimilarities in a test phase.
The novelty of our approach lies in its reformulation for general dissimilarity measures,
which, we think, is more natural for class descriptors. Since dissimilarity measures might
be unbounded, we have also proposed to transform dissimilarities by the sigmoid function,
which makes the LPDD more robust against objects with large dissimilarities. We emphasized the possibility of using the LP procedures for rectangular dissimilarity/similarity
representations, which is especially useful when (dis)similarities are costly to compute.
The LPDD is applied to artificial and real-world datasets and compared to the LPSD detector as proposed in [3]. For all considered datasets, the LPDD yields a more compact target
description than the LPSD. The LPDD is more robust against outliers in the training set, in
particular, when only some objects are considered for a reduced representation. Moreover,
with a proper scaling parameter s of the sigmoid function, the support objects in the LPDD
do not contain outliers, while it seems difficult for the LPSD to achieve the same. In the
original formulation, the support objects of the LPSD tend to lie on the boundary, while
for the LPDD, they are mostly ?inside? the boundary. This means that a removal of such an
object will not impose a drastic change of the LPDD.
In summary, our LPDD method can be recommended when the failure to detect outliers is
more expensive than the costs of a false alarm. It is also possible to enlarge the description
of the LPDD by adding a small constant to ?. Such a constant should be related to the
dissimilarity values in the neighborhood of the boundary. How to choose it, remains an
open issue for further research.
Acknowledgements. This work is partly supported by the Dutch Organization for Scientific Research (NWO) and the European Community Marie Curie Fellowship. The authors are solely responsible for information communicated and the European Commission is not responsible for any views
or results expressed.
References
[1] K.P. Bennett and O.L. Mangasarian. Combining support vector and mathematical programming
methods for induction. In B. Sch?lkopf, C.J.C. Burges, and A.J. Smola, editors, Advances in
Kernel Methods, Support Vector Learning, pages 307?326. MIT Press, Cambridge, MA, 1999.
[2] C. Berg, J.P.R. Christensen, and P. Ressel. Harmonic Analysis on Semigroups. Springer-Verlag,
1984.
[3] C. Campbell and K.P. Bennett. A linear programming approach to novelty detection. In Neural
Information Processing Systems, pages 395?401, 2000.
[4] M.P. Dubuisson and A.K. Jain. Modified Hausdorff distance for object matching. In 12th
Internat. Conference on Pattern Recognition, volume 1, pages 566?568, 1994.
[5] R.P.W. Duin. Compactness and complexity of pattern recognition problems. In Internat. Symposium on Pattern Recognition ?In Memoriam Pierre Devijver?, pages 124?128, Royal Military
Academy, Brussels, 1999.
[6] R.P.W. Duin and E. P?ekalska. Complexity of dissimilarity based pattern classes. In Scandinavian Conference on Image Analysis, 2001.
[7] D.W. Jacobs, D. Weinshall, and Y. Gdalyahu. Classification with non-metric distances: Image
retrieval and class representation. IEEE Trans. on PAMI, 22(6):583?600, 2000.
[8] A.K. Jain and D. Zongker. Representation and recognition of handwritten digits using deformable templates. IEEE Trans. on PAMI, 19(12):1386?1391, 1997.
[9] Mangasarian O.L. Arbitrary-norm separating plane. Operations Research Letters, 24(1-2):15?
23, 1999.
[10] E. P?ekalska, P. Paclik, and R.P.W. Duin. A generalized kernel approach to dissimilarity-based
classification. Journal of Machine Learning Research, 2(2):175?211, 2001.
[11] B. Sch?lkopf, J.C. Platt, A.J. Smola, and R.C. Williamson. Estimating the support of a highdimensional distribution. Neural Computation, 13:1443?1471, 2001.
[12] B. Sch?lkopf, Williamson R.C., A.J. Smola, J. Shawe-Taylor, and J.C. Platt. Support vector
method for novelty detection. In Neural Information Processing Systems, 2000.
[13] D.M.J. Tax. One-class classification. PhD thesis, Delft University of Technology, The Netherlands, 2001.
[14] D.M.J. Tax and R.P.W. Duin. Support vector data description. Machine Learning, 2002. accepted.
[15] V. Vapnik. The Nature of Statistical Learning. Springer, N.Y., 1995.
[16] http://www.sidanet.org.
| 2163 |@word dubuisson:1 norm:2 seems:1 open:2 jacob:1 thereby:2 contains:6 ida:1 shape:2 plot:5 selected:1 plane:1 xk:1 accepting:1 hypersphere:1 provides:1 org:1 simpler:1 five:1 unbounded:3 mathematical:2 constructed:1 become:2 above1:1 symposium:1 consists:2 inside:3 expected:2 behavior:3 p1:1 themselves:1 globally:1 spherical:1 little:1 str:1 equipped:1 becomes:1 estimating:1 bounded:5 moreover:1 weinshall:1 kind:1 string:1 transformation:3 classifier:12 rm:1 scaled:1 control:2 platt:2 positive:3 t1:2 local:1 tends:1 solely:1 pami:2 might:5 gdalyahu:1 practical:1 responsible:2 lost:1 block:1 definite:2 communicated:1 digit:1 procedure:2 area:3 significantly:1 printed:1 convenient:2 matching:1 pre:1 refers:3 convenience:1 close:5 judged:1 nb:3 influence:1 www:1 straightforward:1 attention:1 go:1 worn:1 rectangular:3 simplicity:1 kjaer:1 target:13 construction:1 damaged:1 anomaly:2 programming:10 hypothesis:1 origin:10 element:2 recognition:5 satisfying:1 expensive:1 cut:1 bottom:2 observed:2 solved:1 wj:8 remote:3 trade:1 transforming:1 complexity:2 trained:4 rewrite:1 tight:2 ror:1 basis:1 triangle:4 completely:1 represented:7 sigm:2 train:1 jain:2 fast:1 describe:1 detected:2 approached:1 artificial:7 analyser:1 outside:3 neighborhood:2 say:1 otherwise:1 unseen:1 think:1 transform:2 propose:3 product:1 combining:1 achieve:1 deformable:1 academy:1 tax:2 description:10 cluster:4 requirement:1 optimum:1 object:50 illustrate:1 come:1 implies:1 indicate:1 differ:1 attribute:1 require:2 proposition:1 tighter:1 extension:3 hold:2 proximity:10 around:2 considered:4 lying:1 normal:2 mapping:1 scope:1 adopt:1 smallest:1 nwo:1 vibration:1 city:1 minimization:3 mit:1 gaussian:1 modified:2 avoid:2 pn:6 ax:1 l0:5 detect:2 el:1 attract:1 compactness:2 germany:1 issue:1 classification:3 flexible:1 genuine:1 construct:1 enlarge:1 identical:1 unnecessarily:1 look:1 simplex:3 minimized:1 contaminated:1 t2:3 randomly:2 cheaper:1 replaced:1 geometry:1 delft:3 consisting:1 phase:1 semigroups:1 n1:1 detection:11 organization:1 interest:3 acceptance:2 possibility:1 nl:1 diagnostics:2 yielding:1 unless:1 euclidean:8 taylor:1 theoretical:1 minimal:2 monotonous:1 instance:5 increased:1 soft:3 military:1 cover:1 kekul:1 cost:3 vertex:3 rare:1 uniform:2 conducted:1 commission:1 discriminating:1 off:1 together:2 analogously:1 thesis:1 choose:2 possibly:1 positivity:1 de:1 summarized:1 bold:1 race:1 performed:1 view:2 lot:1 try:2 start:2 parallel:1 slope:1 curie:1 minimize:3 square:5 descriptor:4 characteristic:1 yield:4 t3:3 lkopf:3 identification:2 handwritten:1 monitoring:1 explain:1 detector:1 definition:1 against:6 failure:3 obvious:1 dm:6 naturally:1 proof:3 dataset:5 knowledge:2 dimensionality:1 cj:1 nonmetric:1 campbell:3 higher:2 originally:1 formulation:9 anywhere:1 just:2 smola:3 until:1 d:2 hand:1 nonlinear:2 cage:2 lack:2 continuity:1 indicated:1 scientific:1 building:1 concept:2 contain:1 hausdorff:1 symmetric:1 conditionally:1 generalized:1 tn:1 performs:1 l1:4 reasoning:3 image:3 meaning:1 wise:1 novel:1 recently:2 mangasarian:2 harmonic:1 sigmoid:4 volume:8 discussed:1 belong:2 illness:2 interpretation:1 refer:4 measurement:1 cambridge:1 shawe:1 scandinavian:1 zongker:1 similarity:11 internat:2 etc:1 base:1 belongs:1 reverse:2 verlag:1 inequality:3 prism:6 fault:9 yi:1 seen:1 somewhat:1 impose:2 recognized:1 novelty:4 recommended:2 dashed:2 signal:1 d0:1 characterized:2 calculation:1 retrieval:1 equally:1 basic:1 vision:1 metric:9 dutch:1 kernel:10 represent:1 fellowship:1 interval:1 median:2 concluded:1 crucial:1 sch:3 strict:1 tend:2 member:2 contrary:1 spirit:1 call:1 xj:3 identified:1 inner:1 prototype:2 knowing:1 whether:1 ignored:1 useful:1 proportionally:1 clear:1 netherlands:2 locally:1 ph:1 category:1 reduced:13 http:1 misclassifying:1 fulfilled:1 deteriorates:1 discrete:1 express:1 four:4 reformulation:1 threshold:2 pj:1 marie:1 graph:1 fraction:3 sum:1 letter:1 separation:1 decision:1 scaling:7 comparable:1 abnormal:2 bound:3 quadratic:1 badly:2 duin:4 constraint:1 fourier:1 min:5 brussels:1 ball:4 describes:1 slightly:4 remain:1 smaller:1 wi:3 lp:12 making:1 vji:1 christensen:1 outlier:21 intuitively:1 restricted:1 taken:1 remains:1 describing:2 discus:1 needed:2 know:1 drastic:1 available:1 operation:1 observe:1 pierre:1 robustness:1 original:2 include:1 especially:2 hypercube:1 objective:1 already:2 realized:3 damage:1 costly:2 distance:18 separate:1 mapped:1 berlin:1 separating:1 outer:1 ressel:1 reason:1 induction:1 assuming:1 illustration:3 minimizing:3 difficult:1 mostly:2 robert:1 expense:1 negative:2 stated:1 proper:2 perform:1 upper:3 datasets:4 situation:1 defining:1 extended:1 rn:1 arbitrary:2 community:1 david:1 specified:1 optimized:1 trans:2 address:2 beyond:1 below:1 pattern:4 tudelft:1 built:3 max:1 royal:1 suitable:2 ela:1 natural:4 difficulty:1 indicator:1 technology:2 picture:1 ekalska:2 created:1 fraunhofer:2 acknowledgement:1 tangent:1 removal:1 determining:3 validation:3 sufficient:1 consistent:1 principle:3 dq:1 editor:1 pi:9 row:2 changed:1 summary:1 placed:1 supported:1 dis:3 burges:1 institute:1 template:1 sparse:1 distributed:3 boundary:10 xn:1 valid:1 world:1 author:1 made:2 far:1 emphasize:1 compact:1 implicitly:1 conceptual:1 assumed:1 unnecessary:1 conclude:1 xi:5 why:1 reality:1 table:2 nature:1 robust:6 symmetry:1 reflectivity:1 bearing:4 williamson:2 necessarily:2 european:2 domain:3 main:1 linearly:1 bounding:1 whole:1 arise:1 alarm:2 x1:1 fig:8 representative:1 explicit:1 lq:1 lie:7 atypical:1 emphasized:1 tection:1 oneclass:1 essential:1 false:2 adding:1 effectively:1 gained:1 vapnik:1 phd:1 dissimilarity:50 t4:5 demand:1 margin:6 intersection:1 expressed:2 paclik:1 springer:2 ch:3 devijver:1 satisfies:1 ma:1 marked:4 consequently:3 rbf:2 towards:1 acceleration:1 clidean:1 occ:7 bennett:4 feasible:1 hard:3 change:3 typical:2 except:1 determined:1 uniformly:3 hyperplane:10 wt:4 called:1 partly:1 accepted:1 highdimensional:1 berg:1 support:14 dissimilar:1 violated:2 |
1,278 | 2,164 | Distance Metric Learning, with Application
to Clustering with Side-Information
Eric P. Xing, Andrew Y. Ng, Michael I. Jordan and Stuart Russell
University of California, Berkeley
Berkeley, CA 94720
epxing,ang,jordan,russell @cs.berkeley.edu
Abstract
Many algorithms rely critically on being given a good metric over their
inputs. For instance, data can often be clustered in many ?plausible?
ways, and if a clustering algorithm such as K-means initially fails to find
one that is meaningful to a user, the only recourse may be for the user to
manually tweak the metric until sufficiently good clusters are found. For
these and other applications requiring good metrics, it is desirable that
we provide a more systematic way for users to indicate what they consider ?similar.? For instance, we may ask them to provide examples. In
this paper, we present an algorithm that, given examples of similar (and,
if desired, dissimilar) pairs of points in , learns a distance metric over
that respects these relationships. Our method is based on posing metric learning as a convex optimization problem, which allows us to give
efficient, local-optima-free algorithms. We also demonstrate empirically
that the learned metrics can be used to significantly improve clustering
performance.
1 Introduction
The performance of many learning and datamining algorithms depend critically on their
being given a good metric over the input space. For instance, K-means, nearest-neighbors
classifiers and kernel algorithms such as SVMs all need to be given good metrics that reflect
reasonably well the important relationships between the data. This problem is particularly
acute in unsupervised settings such as clustering, and is related to the perennial problem of
there often being no ?right? answer for clustering: If three algorithms are used to cluster a
set of documents, and one clusters according to the authorship, another clusters according
to topic, and a third clusters according to writing style, who is to say which is the ?right?
answer? Worse, if an algorithm were to have clustered by topic, and if we instead wanted it
to cluster by writing style, there are relatively few systematic mechanisms for us to convey
this to a clustering algorithm, and we are often left tweaking distance metrics by hand.
In this paper, we are interested in the following problem: Suppose a user indicates that
certain points in an input space (say, ) are considered by them to be ?similar.? Can we
automatically learn a distance metric over that respects these relationships, i.e., one that
assigns small distances between the similar pairs? For instance, in the documents example,
we might hope that, by giving it pairs of documents judged to be written in similar styles,
it would learn to recognize the critical features for determining style.
One important family of algorithms that (implicitly) learn metrics are the unsupervised
ones that take an input dataset, and find an embedding of it in some space. This includes
algorithms such as Multidimensional Scaling (MDS) [2], and Locally Linear Embedding
(LLE) [9]. One feature distinguishing our work from these is that we will learn a full metric
over the input space, rather than focusing only on (finding an embedding for) the points in the training set. Our learned metric thus generalizes more easily to
previously unseen data. More importantly, methods such as LLE and MDS also suffer from
the ?no right answer? problem: For example, if MDS finds an embedding that fails to capture the structure important to a user, it is unclear what systematic corrective actions would
be available. (Similar comments also apply to Principal Components Analysis (PCA) [7].)
As in our motivating clustering example, the methods we propose can also be used in a
pre-processing step to help any of these unsupervised algorithms to find better solutions.
In the supervised learning setting, for instance nearest neighbor classification, numerous
attempts have been made to define or learn either local or global metrics for classification.
In these problems, a clear-cut, supervised criterion?classification error?is available and
can be optimized for. (See also [11], for a different way of supervising clustering.) This
literature is too wide to survey here, but some relevant examples include [10, 5, 3, 6],
and [1] also gives a good overview of some of this work. While these methods often
learn good metrics for classification, it is less clear whether they can be used to learn
good, general metrics for other algorithms such as K-means, particularly if the information
available is less structured than the traditional, homogeneous training sets expected by
them.
In the context of clustering, a promising approach was recently proposed by Wagstaff et
al. [12] for clustering with similarity information. If told that certain pairs are ?similar? or
?dissimilar,? they search for a clustering that puts the similar pairs into the same, and dissimilar pairs into different, clusters. This gives a way of using similarity side-information
to find clusters that reflect a user?s notion of meaningful clusters. But similar to MDS and
LLE, the (?instance-level?) constraints that they use do not generalize to previously unseen
data whose similarity/dissimilarity to the training set is not known. We will later discuss
this work in more detail, and also examine the effects of using the methods we propose in
conjunction with these methods.
2 Learning Distance Metrics
, and are given information that certain
!
and are similar
(1)
#
"#!if between
How can we learn a distance metric
points and that respects this;
Suppose we have some set of points
pairs of them are ?similar?:
specifically, so that ?similar? points end up close to each other?
Consider learning a distance metric of the form
"#!%$ '&( "#)$+*,* -.#/*,* & $+0 1-2#4365 7-.#98
5
5;:=<
(2)
5>$@?
To ensure that this be a metric?satisfying non-negativity and the triangle inequality?
we require that
be positive semi-definite,
.1 Setting
gives Euclidean
distance; if we restrict to be diagonal, this corresponds to learning a metric in which
the different axes are given different ?weights?; more generally, parameterizes a family
of Mahalanobis distances over .2 Learning such a distance metric is also equivalent
to finding a rescaling of a data that replaces each point with
and applying the
5
5
5 ACB
Technically, this also allows pseudometrics, where DFEHGJILKNMPO/QSR does not imply ITQUM .
that, but putting the original dataset through a non-linear basis function V and considering
W GXV/Note
GJI!OY7VZGJMPONO\[L]^GXV/GJI!OY_VZGJMPONO , non-linear distance metrics can also be learned.
1
2
standard Euclidean metric to the rescaled data; this will later be useful in visualizing the
learned metrics.
A simple way of defining a criterion for the desired metric is to demand that
pairs of points
in
have, say, small squared distance between them:
. This is trivially solved with
, which is not
useful, and we add the constraint
to ensure that does not
collapse the dataset into a single point. Here, can be a set of pairs of points known to be
?dissimilar? if such information is explicitly available; otherwise, we may take it to be all
pairs not in . This gives the optimization problem:
B
&
,* *
- *,* &
,* *
-S!!** &
!
5 $ <
5
&
,* *
/-2!'*,* B&
" ,* * L
/-2!'*,* &#$
s.t.
(3)
5 : <8
5
(4)
B
% 5
(5)
The choice of the constant 1 in the right hand side of (4) is arbitrary but not important, and
changing it to any other positive constant results only in being replaced by
. Also,
this problem has an objective that is linear in the parameters , and both of the constraints
are also easily verified to be convex. Thus, the optimization problem is convex, which
enables us to derive efficient, local-minima-free algorithms to solve it.
We also note that, while one might consider various alternatives to (4), ?
? would not be a good choice despite its giving a simple linear constraint. It
would result in always being rank 1 (i.e., the data are always projected onto a line). 3
%
!'*,* B& '
5
& **
-
5
2.1 The case of diagonal
5
5 $)( *,+ 5 C C5 BCB 8 8 8ZC5
In the case that we want to learn a diagonal
an efficient algorithm using the Newton-Raphson method. Define
, we can derive
.
*,*
-2 *,* B& -0/1 +324 .
*,*
-2 *,* &657
5 : < ) is equivalent, up to a
It is straightforward
to show that minimizing - (subject to
5
multiplication of by a positive constant, to solving the original problem (3?5). We can
thus use Newton-Raphson to efficiently optimize - .4
- 5 %$ - 5 C 8 8 8Z5
2.2 The case of full
)$
5
9B
5
5 : <
8 :9<;
In the case of learning a full matrix , the constraint that
becomes slightly trickier
to enforce, and Newton?s method often becomes prohibitively expensive (requiring
time to invert the Hessian over
parameters). Using gradient descent and the idea of
iterative projections (e.g., [8]) we derive a different algorithm for this setting.
G>=@?BA C C A DFDFE EGIHBH IKJ"Y IL HBH EM ONO=@?BA C A PD:EQRHBH IKJ"Y IL HBH EM Q
STU VXW]ZY G N[STUVXWL]ZY Q
Y]\ Q'=^?_A A \ GJKI JHYSIL OCGJI`J YSIL O [G . Decomposing
] as ] Qa=cbJdfehg J g [J (always pos] i R ), this gives = Jg [J Y g J:NO= Jjg [J Y Q g J , which we recognize as a Rayleighsible since a
quotient
whose solution is given by (say) solving the generalized eigenvector problem
G like quantity
Y ge#
Q kKY Q g e for the principal eigenvector, and setting g M Q@lXlXl Q g b QSR .
To ensure that m
] iR , whicheqsisr true iff theeqsdiagonal
] JJ are non-negative, we actually
r , whereelements
t is a step-size parameter optimized via a
replace the Newton update npo
by tun]o
line-search to give the largest downhill step subject to Z
] JJuv.R .
3
The proof is reminiscent of the derivation of Fisher?s linear discriminant. Briefly, consider maximizing
, where
4
Iterate
Iterate
5 $ *P+ & !** 5 -25 ** 5"
5 $ *P+ & !** 5 -25 ** 5 B
5
until converges
5 $ 5
& - 5
until convergence
HBH Y BH H Q G:= = Y J M L O e M
J L
Figure 1: Gradient ascent + Iterative projection algorithm. Here,
).
matrices (
*&
HBHhHBH
is the Frobenius norm on
- 5)$
,* * L
!'*,* &
5 )$
**
*,* B&
We pose the equivalent problem:
s.t.
(6)
5 : <8
5
We will use a gradient ascent step on -
(7)
(8)
to optimize (6), followed by the method of
iterative projections to ensure that the
constraints
(7) and (8) hold. Specifically, we will
repeatedly
take
a
gradient
step
, and then repeatedly project into
and
. This gives the
the sets
algorithm shown in Figure 1.5
The motivation
for the specific choice of the problem formulation (6?8) is that projecting
onto or can be done
Specifically, the first projection step
" #inexpensively.
!
involves minimizing a quadratic objective subject to
a single linear constraint; the solution to this is easily found by solving
time)
(in
of
a sparse system of linear equations. The second projection step onto , the space
%$ '& $ all
positive-semi
definite
matrices,
is
done
by
first
finding
the
diagonalization
,
&
(
(
is a diagonal matrix of ?s eigenvalues
and
the& columns
of
where
$
+
,
$
$
?s corresponding
eigenvectors, and taking
, where
*) contains
&
(
(
!
. (E.g., see [4].)
5 $ 5
$ 5
,* * L
"-U'** B&
5
B
* + & *,* 5 - 5 ** B 5 (
$ ( *,+ 8 8 8"
5
$ ( *,+ *
< 8 8 8Z
<
& - 5
B
5
5
B $ 5 5 : <
5 $
B
8 F9
5 $
3
5 $
3
3 Experiments and Examples
We begin by giving some examples of distance metrics learned on artificial data, and then
show how our methods can be used to improve clustering performance.
3.1 Examples of learned distance metrics
Consider the data shown in Figure 2(a), which is divided into two classes (shown by the
different symbols and, where available, colors). Suppose that points in each class are ?similar? to each other, and we are given reflecting this.6 Depending on whether we learn a
diagonal or a full , we obtain:
;<>=
;<
.0/ 12354016 879 1.036 :
?A@06 6 879 3.245 3.286 0.081
:
:
1.007 :
3.286 3.327 0.082
:
:
:
0.081 0.082 0.002
CB
To visualize this, we can use the fact discussed earlier that learning
is equivalent
to finding a rescaling of the data
, that hopefully ?moves? the similar pairs
&
5
&
*,* *,* &
5 ACB
q Er
q E
5
The algorithm shown in the figure includes a small refinement that the gradient step is taken the
direction of the projection of
onto the orthogonal subspace of 'D , so that it will ?minimally?
disrupt the constraint E . Empirically, this modification often significantly speeds up convergence.
6
In the experiments with synthetic data, F was a randomly sampled 1% of all pairs of similar
points.
e
2?class data (original)
2?class data projection (Newton)
5
5
0
0
5
z
z
z
2?class data projection (IP)
?5
?5
5
0
?5
y
?5
20
5
5
0
0
?5
0
y
x
?5
(a)
?5
5
0
20
0
0
y
x
?20
(b)
?20
x
(c)
]
Figure 2: (a) Original data, with the different classes indicated by the different symbols (and colors, where available). (b) Rescaling of data corresponding to learned diagonal . (c) Rescaling
corresponding to full .
]
3?class data (original)
3?class data projection (Newton)
2
?2
2
0
z
0
z
z
2
3?class data projection (IP)
?2
5
0
?5
y
?5
5
5
0
0
?2
0
y
x
?5
(a)
?5
5
0
2
y
(b)
]
Figure 3: (a) Original data. (b) Rescaling corresponding to learned diagonal
sponding to full .
5 ACB
2
0
x
]
0
?2
?2
x
(c)
. (c) Rescaling corre-
together. Figure 2(b,c) shows the result of plotting
. As we see, the algorithm has
successfully brought together the similar points, while keeping dissimilar ones apart.
Figure 3 shows a similar result for a case of three clusters whose centroids differ only
in the x and y directions. As we see in Figure 3(b), the learned diagonal metric correctly
ignores the z direction. Interestingly, in the case of a full , the algorithm finds a surprising
projection of the data onto a line that still maintains the separation of the clusters well.
5
3.2 Application to clustering
One application of our methods is ?clustering with side information,? in which we learn
a distance metric using similarity information, and cluster data using that metric. Specifically, suppose we are given , and told that each pair
means and belong
to the same cluster. We will consider four algorithms for clustering:
Z
C!
L
*,*
-*,* BB
U
2. Constrained K-means: K-means but subject to points
1. K-means using the default Euclidean metric
between points
to define distortion (and ignoring ).
cluster centroids
and
always being
assigned to the same cluster [12].7
**
"- *,* B&
3. K-means + metric: K-means but with distortion defined using the distance metric
learned from .
4. Constrained K-means + metric: Constrained K-means using the distance metric
learned from .
GJI J KNI L O
IKJ I L
U T
GJIKJ Y O M
This is implemented as the usual K-means, except if
F , then during the step in which
points are assigned to cluster centroids , we assign both and to cluster
. More generally, if we imagine drawing an edge between each pair of points in , then all
the points in each resulting connected component E are constrained to lie in the same cluster, which
we pick to be
.
7
GJI L Y O M
U T
O= A E GJIKJY O M
Porjected 2?class data
10
10
0
0
z
z
Original 2?class data
?10
?10
20
y
20
20
0
?20
?20
y
x
(a)
1.
2.
3.
4.
20
0
0
0
?20
?20
x
(b)
K-means: Accuracy = 0.4975
Constrained K-means: Accuracy = 0.5060
K-means + metric: Accuracy = 1
Constrained K-means + metric: Accuracy = 1
] ?s result is
]
$ 8 8 89 ) be the cluster to which point
is assigned by an automatic clustering
Let % (
algorithm, and let % be some ?correct? or desired clustering
of the data.
Following [?], in
the case of 2-cluster data, we will measure how well the % ?s match the % ?s according to
%
$ % $ %
$ %
$
.
<
8 -
Accuracy
B is the indicator function ( $ , * / $ < ). This is equivalent to
where
L
! drawn randomly from the dataset, our clustering
the
probability that for two points ,
Z
and ! belong to same or different
% agrees
with the ?true? clustering % on whether
Figure 4: (a) Original dataset (b) Data scaled according to learned metric. (
shown, but
gave visually indistinguishable results.)
clusters.8
As a simple example, consider Figure 4, which shows a clustering problem in which the
?true clusters? (indicated by the different symbols/colors in the plot) are distinguished by
their -coordinate, but where the data in its original space seems to cluster much better
according to their -coordinate. As shown by the accuracy scores given in the figure, both
K-means and constrained K-means failed to find good clusterings. But by first learning
a distance metric and then clustering according to that metric, we easily find the correct
clustering separating the true clusters from each other. Figure 5 gives another example
showing similar results.
We also applied our methods to 9 datasets from the UC Irvine repository. Here, the ?true
clustering? is given by the data?s class labels. In each, we ran one experiment using ?little? side-information , and one with ?much? side-information. The results are given in
Figure 6.9
We see that, in almost every problem, using a learned diagonal or full metric leads to
significantly improved performance over naive K-means. In most of the problems, using
a learned metric with constrained K-means (the 5th bar for diagonal , 6th bar for full )
also outperforms using constrained K-means alone (4th bar), sometimes by a very large
#
5
8
5
In the case of many (
) clusters, this evaluation metric tends to give inflated scores since
almost any clustering will correctly predict that most pairs are in different clusters. In this setting,
we therefore modified the measure averaging not only ,
drawn uniformly at random, but from
the same cluster (as determined by ) with chance 0.5, and from different clusters with chance 0.5, so
that ?matches? and ?mis-matches? are given the same weight. All results reported here used K-means
with multiple restarts, and are averages over at least 20 trials (except for wine, 10 trials).
9
F was generated by picking a random subset of all pairs of points sharing the same class . In
the case of ?little? side-information, the size of the subset was chosen so that the resulting number of
resulting connected components
(see footnote 7) would be very roughly 90% of the size of the
original dataset. In the case of ?much? side-information, this was changed to 70%.
!
IJ I L
!J
"$#
Projected data
50
50
0
0
z
z
Original data
?50
?50
50
y
50
50
0
?50
50
0
0
?50
y
x
(a)
1.
2.
3.
4.
0
?50
?50
K-means: Accuracy = 0.4993
Constrained K-means: Accuracy = 0.5701
K-means + metric: Accuracy = 1
Constrained K-means + metric: Accuracy = 1
]
Figure 5: (a) Original dataset (b) Data scaled according to learned metric. (
shown, but
gave visually indistinguishable results.)
Boston housing (N=506, C=3, d=13)
ionosphere (N=351, C=2, d=34)
1
1
0.8
0.8
0.8
0.6
0.6
0.6
0.4
0.4
0.4
0.2
0.2
0.2
0
0
Kc=354
wine (N=168, C=3, d=12)
Kc=269
Kc=187
0
balance (N=625, C=3, d=4)
1
1
1
0.8
0.8
0.6
0.6
0.6
0.4
0.4
0.4
0.2
0.2
0.2
0
0
Kc=127
soy bean (N=47, C=4, d=35)
Kc=548
Kc=400
0
protein (N=116, C=6, d=20)
1
1
0.8
0.8
0.8
0.6
0.6
0.6
0.4
0.4
0.4
0.2
0.2
0.2
0
0
Kc=34
Kc=92
Kc=61
Kc=116
Kc=482
Kc=358
diabetes (N=768, C=2, d=8)
1
Kc=41
Kc=133
breast cancer (N=569, C=2, d=30)
0.8
Kc=153
] ?s result is
Iris plants (N=150, C=3, d=4)
1
Kc=447
x
(b)
0
Kc=694
Kc=611
Figure 6: Clustering accuracy on 9 UCI datasets. In each panel, the six bars on the left correspond to
an experiment with ?little? side-information F , and the six on the right to ?much? side-information.
From left to right, the six bars in each set are respectively K-means, K-means
diagonal metric, K-means full metric, Constrained K-means (C-Kmeans), C-Kmeans diagonal metric, and
C-Kmeans full metric. Also shown are : size of dataset; E : number of classes/clusters; : dimensionality of data;
: mean number of connected components (see footnotes 7, 9). 1 s.e. bars
are also shown.
"$#
D
Performance on Wine dataset
1
0.9
0.9
0.8
0.8
performance
performance
Performance on Protein dataset
1
0.7
0.6
0.6
kmeans
c?kmeans
kmeans + metric (diag A)
c?kmeans + metric (diag A)
kmeans + metric (full A)
c?kmeans + metric (full A)
0.5
0
0.1
kmeans
c?kmeans
kmeans + metric (diag A)
c?kmeans + metric (diag A)
kmeans + metric (full A)
c?kmeans + metric (full A)
0.7
0.5
0.2
0
0.1
ratio of constraints
(a)
0.2
I
ratio of constraints
(b)
Figure 7: Plots of accuracy vs. amount of side-information. Here, the -axis gives the fraction of all
pairs of points in the same class that are randomly sampled to be included in F .
margin. Not surprisingly, we also see that having more side-information in typically
leads to metrics giving better clusterings.
Figure 7 also shows two typical examples of how the quality of the clusterings found increases with the amount of side-information. For some problems (e.g., wine), our algorithm learns good diagonal and full metrics quickly with only a very small amount of
side-information; for some others (e.g., protein), the distance metric, particularly the full
metric, appears harder to learn and provides less benefit over constrained K-means.
4 Conclusions
We have presented an algorithm that, given examples of similar pairs of points in , learns
a distance metric that respects these relationships. Our method is based on posing metric
learning as a convex optimization problem, which allowed us to derive efficient, localoptima free algorithms. We also showed examples of diagonal and full metrics learned
from simple artificial examples, and demonstrated on artificial and on UCI datasets how
our methods can be used to improve clustering performance.
References
[1] C. Atkeson, A. Moore, and S. Schaal. Locally weighted learning. AI Review, 1996.
[2] T. Cox and M. Cox. Multidimensional Scaling. Chapman & Hall, London, 1994.
[3] C. Domeniconi and D. Gunopulos. Adaptive nearest neighbor classification using support vector machines. In Advances in Neural Information Processing Systems 14. MIT Press, 2002.
[4] G. H. Golub and C. F. Van Loan. Matrix Computations. Johns Hopkins Univ. Press, 1996.
[5] T. Hastie and R. Tibshirani. Discriminant adaptive nearest neighbor classification. IEEE Transactions on Pattern Analysis and Machine Learning, 18:607?616, 1996.
[6] T.S. Jaakkola and D. Haussler. Exploiting generative models in discriminaive classifier. In Proc.
of Tenth Conference on Advances in Neural Information Processing Systems, 1999.
[7] I.T. Jolliffe. Principal Component Analysis. Springer-Verlag, New York, 1989.
[8] R. Rockafellar. Convex Analysis. Princeton Univ. Press, 1970.
[9] S.T. Roweis and L.K. Saul. Nonlinear dimensionality reduction by locally linear embedding.
Science 290: 2323-2326.
[10] B. Scholkopf and A. Smola. Learning with Kernels. In Press, 2001.
[11] N. Tishby, F. Pereira, and W. Bialek. The information bottleneck method. In Proc. of the 37th
Allerton Conference on Communication, Control and Computing, 1999.
[12] K. Wagstaff, C. Cardie, S. Rogers, and S. Schroedl. Constrained k-means clustering with background knowledge. In Proc. 18th International Conference on Machine Learning, 2001.
| 2164 |@word trial:2 repository:1 briefly:1 cox:2 norm:1 seems:1 pick:1 harder:1 reduction:1 contains:1 score:2 document:3 interestingly:1 outperforms:1 surprising:1 written:1 reminiscent:1 john:1 enables:1 wanted:1 plot:2 update:1 v:1 alone:1 generative:1 provides:1 allerton:1 vxw:1 scholkopf:1 expected:1 roughly:1 examine:1 automatically:1 little:3 considering:1 becomes:2 project:1 begin:1 panel:1 what:2 eigenvector:2 finding:4 berkeley:3 every:1 multidimensional:2 prohibitively:1 classifier:2 scaled:2 control:1 positive:4 local:3 tends:1 gunopulos:1 despite:1 might:2 minimally:1 collapse:1 definite:2 significantly:3 projection:11 pre:1 tweaking:1 protein:3 onto:5 close:1 judged:1 put:1 context:1 applying:1 writing:2 bh:1 kni:1 optimize:2 equivalent:5 demonstrated:1 maximizing:1 straightforward:1 convex:5 survey:1 assigns:1 haussler:1 importantly:1 embedding:5 notion:1 coordinate:2 imagine:1 suppose:4 user:6 homogeneous:1 distinguishing:1 diabetes:1 satisfying:1 particularly:3 expensive:1 cut:1 solved:1 capture:1 connected:3 russell:2 rescaled:1 thee:1 ran:1 pd:1 depend:1 solving:3 technically:1 eric:1 basis:1 triangle:1 perennial:1 easily:4 po:1 various:1 corrective:1 derivation:1 univ:2 london:1 artificial:3 whose:3 plausible:1 solve:1 say:4 distortion:2 otherwise:1 drawing:1 unseen:2 ip:2 housing:1 eigenvalue:1 propose:2 pseudometrics:1 relevant:1 uci:2 iff:1 roweis:1 frobenius:1 exploiting:1 convergence:2 cluster:29 optimum:1 converges:1 help:1 supervising:1 andrew:1 derive:4 pose:1 depending:1 nearest:4 implemented:1 c:1 quotient:1 indicate:1 involves:1 inflated:1 differ:1 direction:3 correct:2 bean:1 rogers:1 require:1 assign:1 clustered:2 hold:1 sufficiently:1 considered:1 hall:1 visually:2 cb:1 predict:1 visualize:1 wine:4 proc:3 label:1 largest:1 agrees:1 successfully:1 weighted:1 hope:1 brought:1 mit:1 always:4 modified:1 rather:1 jaakkola:1 conjunction:1 ax:1 schaal:1 rank:1 indicates:1 centroid:3 typically:1 initially:1 kc:18 interested:1 classification:6 constrained:14 uc:1 having:1 ng:1 manually:1 chapman:1 stuart:1 unsupervised:3 others:1 few:1 randomly:3 recognize:2 replaced:1 attempt:1 inexpensively:1 evaluation:1 golub:1 edge:1 orthogonal:1 euclidean:3 desired:3 instance:6 column:1 earlier:1 trickier:1 tweak:1 subset:2 too:1 tishby:1 motivating:1 reported:1 acb:3 answer:3 synthetic:1 international:1 systematic:3 told:2 picking:1 michael:1 together:2 quickly:1 hopkins:1 squared:1 reflect:2 worse:1 style:4 rescaling:6 includes:2 rockafellar:1 explicitly:1 later:2 xing:1 maintains:1 accuracy:12 who:1 efficiently:1 correspond:1 generalize:1 critically:2 zy:2 cardie:1 footnote:2 sharing:1 npo:1 proof:1 mi:1 sampled:2 irvine:1 dataset:10 ask:1 color:3 knowledge:1 dimensionality:2 actually:1 reflecting:1 focusing:1 appears:1 supervised:2 restarts:1 improved:1 formulation:1 done:2 smola:1 until:3 hand:2 nonlinear:1 hopefully:1 quality:1 indicated:2 effect:1 requiring:2 true:5 assigned:3 moore:1 mahalanobis:1 visualizing:1 during:1 indistinguishable:2 authorship:1 iris:1 criterion:2 generalized:1 demonstrate:1 recently:1 empirically:2 overview:1 discussed:1 belong:2 ai:1 automatic:1 trivially:1 acute:1 similarity:4 add:1 showed:1 apart:1 certain:3 verlag:1 inequality:1 minimum:1 semi:2 full:18 desirable:1 multiple:1 match:3 raphson:2 divided:1 breast:1 metric:67 kernel:2 sponding:1 sometimes:1 invert:1 background:1 want:1 dfehg:1 ascent:2 comment:1 subject:4 jordan:2 iterate:2 gave:2 hastie:1 restrict:1 idea:1 parameterizes:1 bottleneck:1 whether:3 six:3 pca:1 suffer:1 hessian:1 york:1 action:1 repeatedly:2 generally:2 useful:2 clear:2 eigenvectors:1 amount:3 ang:1 locally:3 svms:1 correctly:2 tibshirani:1 putting:1 four:1 drawn:2 changing:1 verified:1 tenth:1 fraction:1 family:2 almost:2 separation:1 scaling:2 followed:1 corre:1 replaces:1 quadratic:1 constraint:10 speed:1 relatively:1 structured:1 according:8 slightly:1 em:2 modification:1 ikj:3 projecting:1 wagstaff:2 taken:1 recourse:1 equation:1 previously:2 discus:1 mechanism:1 jolliffe:1 ge:1 end:1 generalizes:1 available:6 decomposing:1 apply:1 enforce:1 datamining:1 distinguished:1 alternative:1 original:12 clustering:30 include:1 ensure:4 newton:6 qsr:2 giving:4 objective:2 move:1 schroedl:1 quantity:1 md:4 traditional:1 diagonal:14 unclear:1 usual:1 gradient:5 bialek:1 subspace:1 distance:21 separating:1 topic:2 discriminant:2 relationship:4 gji:4 ratio:2 minimizing:2 balance:1 negative:1 ba:2 datasets:3 descent:1 defining:1 communication:1 arbitrary:1 pair:18 optimized:2 california:1 learned:16 qa:1 bar:6 pattern:1 tun:1 critical:1 rely:1 indicator:1 improve:3 epxing:1 imply:1 numerous:1 axis:1 negativity:1 naive:1 review:1 literature:1 multiplication:1 determining:1 plant:1 plotting:1 cancer:1 changed:1 surprisingly:1 free:3 keeping:1 side:14 lle:3 neighbor:4 wide:1 taking:1 saul:1 sparse:1 benefit:1 van:1 default:1 ignores:1 made:1 c5:2 projected:2 refinement:1 adaptive:2 atkeson:1 transaction:1 bb:1 implicitly:1 global:1 disrupt:1 search:2 iterative:3 promising:1 learn:12 reasonably:1 ca:1 ignoring:1 posing:2 diag:4 motivation:1 allowed:1 convey:1 f9:1 fails:2 downhill:1 pereira:1 lie:1 third:1 learns:3 specific:1 showing:1 er:1 symbol:3 ionosphere:1 diagonalization:1 dissimilarity:1 demand:1 margin:1 boston:1 ysi:1 failed:1 kky:1 springer:1 corresponds:1 chance:2 kmeans:15 replace:1 fisher:1 included:1 specifically:4 except:2 uniformly:1 determined:1 averaging:1 typical:1 loan:1 principal:3 domeniconi:1 meaningful:2 support:1 bcb:1 dissimilar:5 princeton:1 gxv:2 |
1,279 | 2,165 | Charting a Manifold
Matthew Brand
Mitsubishi Electric Research Labs
201 Broadway, Cambridge MA 02139 USA
www.merl.com/people/brand/
Abstract
We construct a nonlinear mapping from a high-dimensional sample space
to a low-dimensional vector space, effectively recovering a Cartesian
coordinate system for the manifold from which the data is sampled.
The mapping preserves local geometric relations in the manifold and is
pseudo-invertible. We show how to estimate the intrinsic dimensionality
of the manifold from samples, decompose the sample data into locally
linear low-dimensional patches, merge these patches into a single lowdimensional coordinate system, and compute forward and reverse mappings between the sample and coordinate spaces. The objective functions
are convex and their solutions are given in closed form.
1
Nonlinear dimensionality reduction (NLDR) by charting
Charting is the problem of assigning a low-dimensional coordinate system to data points
in a high-dimensional sample space. It is presumed that the data lies on or near a lowdimensional manifold embedded in the sample space, and that there exists a 1-to-1 smooth
nonlinear transform between the manifold and a low-dimensional vector space. The datamodeler?s goal is to estimate smooth continuous mappings between the sample and coordinate spaces. Often this analysis will shed light on the intrinsic variables of the datagenerating phenomenon, for example, revealing perceptual or configuration spaces.
Our goal is to find a mapping?expressed as a kernel-based mixture of linear projections?
that minimizes information loss about the density and relative locations of sample points.
This constraint is expressed in a posterior that combines a standard gaussian mixture model
(GMM) likelihood function with a prior that penalizes uncertainty due to inconsistent projections in the mixture. Section 3 develops a special case where this posterior is unimodal
and maximizable in closed form, yielding a GMM whose covariances reveal a patchwork of
overlapping locally linear subspaces that cover the manifold. Section 4 shows that for this
(or any) GMM and a choice of reduced dimension d, there is a unique, closed-form solution
for a minimally distorting merger of the subspaces into a d-dimensional coordinate space,
as well as an reverse mapping defining the surface of the manifold in the sample space.
The intrinsic dimensionality d of the data manifold can be estimated from the growth process of point-to-point distances. In analogy to differential geometry, we call the subspaces
?charts? and their merger the ?connection.? Section 5 considers example problems where
these methods are used to untie knots, unroll and untwist sheets, and visualize video data.
1.1 Background
Topology-neutral
NLDR
algorithms can be divided into those that compute mappings, and
those that directly compute low-dimensional embeddings. The field has its roots in mapping algorithms: DeMers and Cottrell [3] proposed using auto-encoding neural networks
with a hidden layer ? bottleneck,? effectively casting dimensionality reduction as a compression problem. Hastie defined principal curves [5] as nonparametric 1 D curves that pass
through the center of ? nearby? data points. A rich literature has grown up around properly
regularizing this approach and extending it to surfaces. Smola and colleagues [10] analyzed
the NLDR problem in the broader framework of regularized quantization methods.
More recent advances aim for embeddings: Gomes and Mojsilovic [4] treat manifold completion as an anisotropic diffusion problem, iteratively expanding points until they connect
to their neighbors. The I SO M AP algorithm [12] represents remote distances as sums of a
trusted set of distances between immediate neighbors, then uses multidimensional scaling
to compute a low-dimensional embedding that minimally distorts all distances. The locally
linear embedding algorithm (LLE) [9] represents each point as a weighted combination of
a trusted set of nearest neighbors, then computes a minimally distorting low-dimensional
barycentric embedding. They have complementary strengths: I SO M AP handles holes well
but can fail if the data hull is nonconvex [12]; and vice versa for LLE [9]. Both offer embeddings without mappings. It has been noted that trusted-set methods are vulnerable to
noise because they consider the subset of point-to-point relationships that has the lowest
signal-to-noise ratio; small changes to the trusted set can induce large changes in the set of
constraints on the embedding, making solutions unstable [1].
In a return to mapping, Roweis and colleagues [8] proposed global coordination? learning
a mixture of locally linear projections from sample to coordinate space. They constructed
a posterior that penalizes distortions in the mapping, and gave a expectation-maximization
(EM) training rule. Innovative use of variational methods highlighted the difficulty of even
hill-climbing their multimodal posterior. Like [2, 7, 6, 8], the method we develop below is
a decomposition of the manifold into locally linear neighborhoods. It bears closest relation
to global coordination [8], although by a different construction of the problem, we avoid
hill-climbing a spiky posterior and instead develop a closed-form solution.
2
Estimating locally linear scale and intrinsic dimensionality
.
We begin with matrix of sample points Y = [y1 , ? ? ? , yN ], yn ? RD populating a Ddimensional sample space, and a conjecture that these points are samples from a manifold M of intrinsic dimensionality d < D. We seek a mapping onto a vector space
.
G(Y) ? X = [x1 , ? ? ? , xN ], xn ? Rd and 1-to-1 reverse mapping G?1 (X) ? Y such that
local relations between nearby points are preserved (this will be formalized below). The
map G should be non-catastrophic, that is, without folds: Parallel lines on the manifold in
RD should map to continuous smooth non-intersecting curves in Rd . This guarantees that
linear operations on X such as interpolation will have reasonable analogues on Y.
Smoothness means that at some scale r the mapping from a neighborhood on M to Rd is
effectively linear. Consider a ball of radius r centered on a data point and containing n(r)
data points. The count n(r) grows as rd , but only at the locally linear scale; the grow rate
is inflated by isotropic noise at smaller scales and by embedding curvature at larger scales.
.
To estimate r, we look at how the r-ball grows as points are added to it, tracking c(r) =
d
d log n(r) log r. At noise scales, c(r) ? 1/D < 1/d, because noise has distributed points in
all directions with equal probability. At the scale at which curvature becomes significant,
c(r) < 1/d, because the manifold is no longer perpendicular to the surface of the ball, so
the ball does not have to grow as fast to accommodate new points. At the locally linear
scale, the process peaks at c(r) = 1/d, because points are distributed only in the directions
of the manifold?s local tangent space. The maximum of c(r) therefore gives an estimate
of both the scale and the local dimensionality of the manifold (see figure 1), provided that
the ball hasn?t expanded to a manifold boundary? boundaries have lower dimension than
Scale behavior of a 1D manifold in 2-space
Point?count growth process on a 2D manifold in 3?space
1
10
radial growth process
1D hypothesis
2D hypothesis
3D hypothesis
radius (log scale)
samples
noise scale
locally linear scale
curvature scale
0
10
2
1
10
2
10
#points (log scale)
3
10
Figure 1: Point growth processes. L EFT: At the locally linear scale, the number of points
in an r-ball grows as rd ; at noise and curvature scales it grows faster. R IGHT: Using the
point-count growth process to find the intrinsic dimensionality of a 2D manifold nonlinearly
embedded in 3-space (see figure 2). Lines of slope 1/3 , 1/2 , and 1 are fitted to sections of the
log r/ log nr curve. For neighborhoods of radius r ? 1 with roughly n ? 10 points, the slope
peaks at 1/2 indicating a dimensionality of d = 2. Below that, the data appears 3 D because
it is dominated by noise (except for n ? D points); above, the data appears >2 D because of
manifold curvature. As the r-ball expands to cover the entire data-set the dimensionality
appears to drop to 1 as the process begins to track the 1D edges of the 2D sheet.
the manifold. For low-dimensional manifolds such as sheets, the boundary submanifolds
(edges and corners) are very small relative to the full manifold, so the boundary effect is
typically limited to a small rise in c(r) as r approaches the scale of the entire data set. In
practice, our code simply expands an r-ball at every point and looks for the first peak in
c(r), averaged over many nearby r-balls. One can estimate d and r globally or per-point.
3
Charting the data
In the charting step we find a soft partitioning of the data into locally linear low-dimensional
neighborhoods, as a prelude to computing the connection that gives the global lowdimensional embedding. To minimize information loss in the connection, we require that
the data points project into a subspace associated with each neighborhood with (1) minimal
loss of local variance and (2) maximal agreement of the projections of nearby points into
nearby neighborhoods. Criterion (1) is served by maximizing the likelihood function of a
Gaussian mixture model (GMM) density fitted to the data:
.
p(yi |?, ?) = ? j p(yi |? j , ? j ) p j = ? j N (yi ; ? j , ? j ) p j .
(1)
Each gaussian component defines a local neighborhood centered around ? j with axes defined by the eigenvectors of ? j . The amount of data variance along each axis is indicated
by the eigenvalues of ? j ; if the data manifold is locally linear in the vicinity of the ? j , all
but the d dominant eigenvalues will be near-zero, implying that the associated eigenvectors constitute the optimal variance-preserving local coordinate system. To some degree
likelihood maximization will naturally realize this property: It requires that the GMM components shrink in volume to fit the data as tightly as possible, which is best achieved by
positioning the components so that they ? pancake? onto locally flat collections of datapoints. However, this state of affairs is easily violated by degenerate (zero-variance) GMM
components or components fitted to overly small enough locales where the data density off
the manifold is comparable to density on the manifold (e.g., at the noise scale). Consequently a prior is needed.
Criterion (2) implies that neighboring partitions should have dominant axes that span similar subspaces, since disagreement (large subspace angles) would lead to inconsistent projections of a point and therefore uncertainty about its location in a low-dimensional coordinate space. The principal insight is that criterion (2) is exactly the cost of coding the
location of a point in one neighborhood when it is generated by another neighborhood? the
cross-entropy between the gaussian models defining the two neighborhoods:
D(N1 kN2 )
=
=
Z
dy N (y; ?1 ,?1 ) log
N (y; ?1 ,?1 )
N (y; ?2 ,?2 )
> ?1
?1
(log |??1
1 ?2 | + trace(?2 ?1 ) + (?2 ??1 ) ?2 (?2 ??1 ) ? D)/2. (2)
Roughly speaking, the terms in (2) measure differences in size, orientation, and position,
respectively, of two coordinate frames located at the means ?1 , ?2 with axes specified by
the eigenvectors of ?1 , ?2 . All three terms decline to zero as the overlap between the two
frames is maximized. To maximize consistency between adjacent neighborhoods, we form
.
the prior p(?, ?) = exp[? ?i6= j mi (? j )D(Ni kN j )], where mi (? j ) is a measure of co-locality.
Unlike global coordination [8], we are not asking that the dominant axes in neighboring
charts are aligned? only that they span nearly the same subspace. This is a much easier
objective to satisfy, and it contains a useful special case where the posterior p(?, ?|Y) ?
?i p(yi |?, ?)p(?, ?) is unimodal and can be maximized in closed form: Let us associate a
gaussian neighborhood with each data-point, setting ?i = yi ; take all neighborhoods to be
a priori equally probable, setting pi = 1/N; and let the co-locality measure be determined
from some local kernel. For example, in this paper we use mi (? j ) ? N (? j ; ?i , ?2 ), with
the scale parameter ? specifying the expected size of a neighborhood on the manifold in
sample space. A reasonable choice is ? = r/2, so that 2erf(2) > 99.5% of the density of
mi (? j ) is contained in the area around yi where the manifold is expected to be locally linear.
With uniform pi and ?i , mi (? j ) and fixed, the MAP estimates of the GMM covariances are
!,
?i =
? mi (? j ) (y j ? ?i )(y j ? ?i )> + (? j ? ?i )(? j ? ?i )> + ? j
? mi (? j ) (3).
j
j
Note that each covariance ?i is dependent on all other ? j . The MAP estimators for all
covariances can be arranged into a set of fully constrained linear equations and solved exactly for their mutually optimal values. This key step brings nonlocal information about
the manifold?s shape into the local description of each neighborhood, ensuring that adjoining neighborhoods have similar covariances and small angles between their respective
subspaces. Even if a local subset of data points are dense in a direction perpendicular to
the manifold, the prior encourages the local chart to orient parallel to the manifold as part
of a globally optimal solution, protecting against a pathology noted in [8]. Equation (3) is
easily adapted to give a reduced number of charts and/or charts centered on local centroids.
4
Connecting the charts
We now build a connection for set of charts specified as an arbitrary nondegenerate GMM. A
GMM gives a soft partitioning of the dataset into neighborhoods of mean ?k and covariance
?k . The optimal variance-preserving low-dimensional coordinate system for each neighborhood derives from its weighted principal component analysis, which is exactly specified
by the eigenvectors of its covariance matrix: Eigendecompose Vk ?k V>
k ? ?k with eigen.
values in descending order on the diagonal of ?k and let Wk = [Id , 0]V>
the operator
k be
.
projecting points into the kth local chart, such that local chart coordinate uki = Wk (yi ? ?k )
.
and Uk = [uk1 , ? ? ? , ukN ] holds the local coordinates of all points.
Our goal is to sew together all charts into a globally consistent low-dimensional coordinate
system. For each chart there will be a low-dimensional affine transform Gk ? R(d+1)?d
that projects Uk into the global coordinate space. Summing over all charts, the weighted
average of the projections of point yi into the low-dimensional vector space is
W j (y ? ? j )
u ji
.
.
d
c
p j|y (y)
?
xi |yi = ? G j
x|y = ? G j
p j|y (yi ),
(4)
1
1
j
j
where pk|y (y) ? pk N (y; ?k , ?k ), ?k pk|y (y) = 1 is the probability that chart k generates
point y. As pointed out in [8], if a point has nonzero probabilities in two charts, then there
should be affine transforms of those two charts that map the point to the same place in a
global coordinate space. We set this up as a weighted least-squares problem:
2
uki
u ji
.
.
G
G = [G1 , ? ? ? , GK ] = arg min ? pk|y (yi )p j|y (yi )
?
G
(5)
j
k
1
1
F
Gk ,G j i
Equation (5) generates a homogeneous set of equations that determines a solution up to an
affine transform of G. There are two solution methods. First, let us temporarily anchor one
neighborhood at the origin to fix this indeterminacy. This adds the constraint G1 = [I, 0]> .
.
To solve, define indicator matrix Fk = [0, ? ? ? , 0, I, 0, ? ? ? , 0]> with the identity ma.
trix occupying the kth block, such that Gk = GFk . Let the diagonal of Pk =
diag([pk|y (y1 ), ? ? ? , pk|y (yN )]) record the per-point posteriors of chart k. The squared error
of the connection is then a sum of of all patch-to-anchor and patch-to-patch inconsistencies:
"
#
2
2
U
U
.
.
1
E = ?
(GUk ? 0 )Pk P1
+ ?
(GU j ? GUk )P j Pk
F ; Uk = Fk 1k .
F
j6=k
k
(6)
Setting dE /dG = 0 and solving to minimize convex E gives
!
!?1
>
G =
?
Uk P2k
k
?
j6=k
P2j
U>
k ?
?
j6=k
Uk P2k P2j U>j
?
k
Uk P2k P21
U1
0
> !
.
(7)
We now remove the dependence on a reference neighborhood G1 by rewriting equation 5,
G = arg min ? j6=k k(GU j ? GUk )P j Pk k2F = kGQk2F = trace(GQQ> G> ) ,
(8)
G
.
where Q = ? j6=k U j ? Uk P j Pk . If we require that GG> = I to prevent degenerate
solutions, then equation (8) is solved (up to rotation in coordinate space) by setting G> to
the eigenvectors associated with the smallest eigenvalues of QQ> . The eigenvectors can be
computed efficiently without explicitly forming QQ> ; other numerical efficiencies obtain
by zeroing any vanishingly small probabilities in each Pk , yielding a sparse eigenproblem.
A more interesting strategy is to numerically condition the problem by calculating the
trailing eigenvectors of QQ> + 1. It can be shown that this maximizes the posterior
2
p(G|Q) ? p(Q|G)p(G) ? e?kGQkF e?kG1k , where the prior p(G) favors a mapping G
whose unit-norm rows are also zero-mean. This maximizes variance in each row of G
and thereby spreads the projected points broadly and evenly over coordinate space.
The solutions for MAP charts (equation (5)) and connection (equation (8)) can be applied
to any well-fitted mixture of gaussians/factors1 /PCAs density model; thus large eigenproblems can be avoided by connecting just a small number of charts that cover the data.
1 We
thank reviewers for calling our attention to Teh & Roweis ([11]? in this volume), which
shows how to connect a set of given local dimensionality reducers in a generalized eigenvalue problem that is related to equation (8).
charting
(projection onto coordinate space)
charting
best Isomap
LLE, n=5
LLE, n=6
LLE, n=7
LLE, n=8
LLE, n=9
LLE, n=10
XZ view
random subset
of local charts
XYZ view
data (linked)
embedding, XY view
XY view
original data
reconstruction
(back?projected coordinate grid)
best LLE
(regularized)
Figure 2: The twisted curl problem. L EFT: Comparison of charting, I SO M AP, & LLE.
400 points are randomly sampled from the manifold with noise. Charting is the only
method that recovers the original space without catastrophes (folding), albeit with some
shear. R IGHT: The manifold is regularly sampled (with noise) to illustrate the forward
and backward projections. Samples are shown linked into lines to help visualize the manifold structure. Coordinate axes of a random selection of charts are shown as bold lines.
Connecting subsets of charts such as this will also give good mappings. The upper right
quadrant shows various LLE results. At bottom we show the charting solution and the
reconstructed (back-projected) manifold, which smooths out the noise.
Once the connection is solved, equation (4) gives the forward projection of any point y
down into coordinate space. There are several numerically distinct candidates for the backprojection: posterior mean, mode, or exact inverse. In general, there may not be a unique
posterior mode and the exact inverse is not solvable in closed form (this is also true of [8]).
Note that chart-wise projection defines a complementary density in coordinate space
0
[Id , 0]?k [Id , 0]> 0
, Gk
px|k (x) = N (x; Gk
G>
(9)
k ).
1
0
0
Let p(y|x, k), used to map x into subspace k on the surface of the manifold, be a Dirac delta
function whose mean is a linear function of x. Then the posterior mean back-projection is
obtained by integrating out uncertainty over which chart generates x:
+
!
I
0
>
c
y|x = ? pk|x (x) ?k + Wk Gk
x ? Gk
,
(10)
0
1
k
(?)+
where
denotes pseudo-inverse. In general, a back-projecting map should not reconstruct the original points. Instead, equation (10) generates a surface that passes through the
weighted average of the ?i of all the neighborhoods in which yi has nonzero probability,
much like a principal curve passes through the center of each local group of points.
5
Experiments
Synthetic examples: 400 2 D points were randomly sampled from a 2 D square and embedded in 3 D via a curl and twist, then contaminated with gaussian noise. Even if noiselessly
sampled, this manifold cannot be ? unrolled? without distortion. In addition, the outer curl
is sampled much less densely than the inner curl. With an order of magnitude fewer points,
higher noise levels, no possibility of an isometric mapping, and uneven sampling, this is
arguably a much more challenging problem than the ? swiss roll? and ? s-curve? problems
featured in [12, 9, 8, 1]. Figure 2LEFT contrasts the (unique) output of charting and the
best outputs obtained from I SO M AP and LLE (considering all neighborhood sizes between
2 and 20 points). I SO M AP and LLE show catastrophic folding; we had to change LLE?s
b. data, yz view
c. local charts
d. 2D embedding
e. 1D embedding
1D ordinate
a. data, xy view
true manifold arc length
Figure 3: Untying a trefoil knot ( ) by charting. 900 noisy samples from a 3 D-embedded
1 D manifold are shown as connected dots in front (a) and side (b) views. A subset of charts
is shown in (c). Solving for the 2 D connection gives the ? unknot? in (d). After removing
some points to cut the knot, charting gives a 1 D embedding which we plot against true
manifold arc length in (e); monotonicity (modulo noise) indicates correctness.
Three principal degrees of freedom recovered from raw jittered images
pose
scale
expression
images synthesized via backprojection of straight lines in coordinate space
Figure 4: Modeling the manifold of facial images from raw video. Each row contains
images synthesized by back-projecting an axis-parallel straight line in coordinate space
onto the manifold in image space. Blurry images correspond to points on the manifold
whose neighborhoods contain few if any nearby data points.
regularization in order to coax out nondegenerate (>1 D) solutions. Although charting is
not designed for isometry, after affine transform the forward-projected points disagree with
the original points with an RMS error of only 1.0429, lower than the best LLE (3.1423) or
best I SO M AP (1.1424, not shown). Figure 2RIGHT shows the same problem where points
are sampled regularly from a grid, with noise added before and after embedding. Figure 3
shows a similar treatment of a 1 D line that was threaded into a 3 D trefoil knot, contaminated
with gaussian noise, and then ? untied? via charting.
Video: We obtained a 1965-frame video sequence (courtesy S. Roweis and B. Frey) of
20 ? 28-pixel images in which B.F. strikes a variety of poses and expressions. The video
is heavily contaminated with synthetic camera jitters. We used raw images, though image
processing could have removed this and other uninteresting sources of variation. We took a
500-frame subsequence and left-right mirrored it to obtain 1000 points in 20 ? 28 = 560D
image space. The point-growth process peaked just above d = 3 dimensions. We solved for
25 charts, each centered on a random point, and a 3D connection. The recovered degrees
of freedom? recognizable as pose, scale, and expression? are visualized in figure 4.
original data
stereographic map to 3D fishbowl
charting
Figure 5: Flattening a fishbowl. From the left: Original 2000?2D points; their stereographic mapping to a 3D fishbowl; its 2D embedding recovered using 500 charts; and the
stereographic map. Fewer charts lead to isometric mappings that fold the bowl (not shown).
Conformality: Some manifolds can be flattened conformally (preserving local angles) but
not isometrically. Figure 5 shows that if the data is finely charted, the connection behaves
more conformally than isometrically. This problem was suggested by J. Tenenbaum.
6
Discussion
Charting breaks kernel-based NLDR into two subproblems: (1) Finding a set of datacovering locally linear neighborhoods (? charts? ) such that adjoining neighborhoods span
maximally similar subspaces, and (2) computing a minimal-distortion merger (? connection? ) of all charts. The solution to (1) is optimal w.r.t. the estimated scale of local linearity
r; the solution to (2) is optimal w.r.t. the solution to (1) and the desired dimensionality d.
Both problems have Bayesian settings. By offloading the nonlinearity onto the kernels,
we obtain least-squares problems and closed form solutions. This scheme is also attractive
because large eigenproblems can be avoided by using a reduced set of charts.
The dependence on r, like trusted-set methods, is a potential source of solution instability. In practice the point-growth estimate seems fairly robust to data perturbations (to be
expected if the data density changes slowly over a manifold of integral Hausdorff dimension), while the use of a soft neighborhood partitioning appears to make charting solutions
reasonably stable to variations in r. Eigenvalue stability analyses may prove useful here.
Ultimately, we would prefer to integrate r out. In contrast, use of d appears to be a virtue:
Unlike other eigenvector-based methods, the best d-dimensional embedding is not merely
a linear projection of the best d + 1-dimensional embedding; a unique distortion is found
for each value of d that maximizes the information content of its embedding.
Why does charting performs well on datasets where the signal-to-noise ratio confounds
recent state-of-the-art methods? Two reasons may be adduced: (1) Nonlocal information
is used to construct both the system of local charts and their global connection. (2) The
mapping only preserves the component of local point-to-point distances that project onto
the manifold; relationships perpendicular to the manifold are discarded. Thus charting uses
global shape information to suppress noise in the constraints that determine the mapping.
Acknowledgments
Thanks to J. Buhmann, S. Makar, S. Roweis, J. Tenenbaum, and anonymous reviewers for
insightful comments and suggested ? challenge? problems.
References
[1] M. Balasubramanian and E. L. Schwartz. The IsoMap algorithm and topological stability. Science, 295(5552):7, January 2002.
[2] C. Bregler and S. Omohundro. Nonlinear image interpolation using manifold learning. In
NIPS?7, 1995.
[3] D. DeMers and G. Cottrell. Nonlinear dimensionality reduction. In NIPS?5, 1993.
[4] J. Gomes and A. Mojsilovic. A variational approach to recovering a manifold from sample
points. In ECCV, 2002.
[5] T. Hastie and W. Stuetzle. Principal curves. J. Am. Statistical Assoc, 84(406):502?516, 1989.
[6] G. Hinton, P. Dayan, and M. Revow. Modeling the manifolds of handwritten digits. IEEE
Trans. Neural Networks, 8, 1997.
[7] N. Kambhatla and T. Leen. Dimensionality reduction by local principal component analysis.
Neural Computation, 9, 1997.
[8] S. Roweis, L. Saul, and G. Hinton. Global coordination of linear models. In NIPS?13, 2002.
[9] S. T. Roweis and L. K. Saul. Nonlinear dimensionality reduction by locally linear embedding.
Science, 290:2323?2326, December 22 2000.
[10] A. Smola, S. Mika, B. Sch?lkopf, and R. Williamson. Regularized principal manifolds. Machine Learning, 1999.
[11] Y. W. Teh and S. T. Roweis. Automatic alignment of hidden representations. In NIPS?15, 2003.
[12] J. B. Tenenbaum, V. de Silva, and J. C. Langford. A global geometric framework for nonlinear
dimensionality reduction. Science, 290:2319?2323, December 22 2000.
| 2165 |@word compression:1 norm:1 seems:1 seek:1 mitsubishi:1 covariance:7 decomposition:1 datagenerating:1 thereby:1 accommodate:1 reduction:6 configuration:1 contains:2 recovered:3 com:1 assigning:1 realize:1 cottrell:2 numerical:1 partition:1 shape:2 remove:1 drop:1 plot:1 designed:1 implying:1 fewer:2 merger:3 isotropic:1 affair:1 record:1 location:3 along:1 constructed:1 differential:1 prove:1 combine:1 recognizable:1 expected:3 presumed:1 roughly:2 p1:1 xz:1 behavior:1 untying:1 globally:3 balasubramanian:1 considering:1 becomes:1 begin:2 estimating:1 provided:1 project:3 maximizes:3 twisted:1 linearity:1 lowest:1 submanifolds:1 minimizes:1 eigenvector:1 finding:1 locale:1 guarantee:1 pseudo:2 every:1 multidimensional:1 expands:2 p2j:2 growth:7 shed:1 isometrically:2 exactly:3 fishbowl:3 assoc:1 schwartz:1 partitioning:3 uk:7 unit:1 yn:3 arguably:1 before:1 local:24 treat:1 frey:1 encoding:1 id:3 interpolation:2 merge:1 ap:6 mika:1 minimally:3 specifying:1 challenging:1 co:2 limited:1 perpendicular:3 averaged:1 unique:4 camera:1 stereographic:3 acknowledgment:1 practice:2 block:1 swiss:1 digit:1 stuetzle:1 area:1 featured:1 revealing:1 projection:12 induce:1 radial:1 quadrant:1 integrating:1 onto:6 cannot:1 selection:1 sheet:3 operator:1 instability:1 descending:1 www:1 map:10 reviewer:2 center:2 maximizing:1 courtesy:1 attention:1 convex:2 formalized:1 rule:1 insight:1 estimator:1 datapoints:1 embedding:16 handle:1 stability:2 coordinate:25 variation:2 qq:3 construction:1 heavily:1 modulo:1 exact:2 homogeneous:1 us:2 hypothesis:3 origin:1 agreement:1 associate:1 located:1 cut:1 bottom:1 solved:4 connected:1 remote:1 removed:1 reducer:1 conformality:1 ultimately:1 solving:2 efficiency:1 gu:2 multimodal:1 easily:2 bowl:1 various:1 grown:1 distinct:1 fast:1 neighborhood:26 whose:4 larger:1 solve:1 distortion:4 reconstruct:1 favor:1 erf:1 g1:3 transform:4 highlighted:1 noisy:1 sequence:1 eigenvalue:5 took:1 reconstruction:1 lowdimensional:3 maximal:1 vanishingly:1 neighboring:2 aligned:1 degenerate:2 roweis:7 description:1 dirac:1 sew:1 extending:1 help:1 illustrate:1 develop:2 completion:1 pose:3 nearest:1 indeterminacy:1 recovering:2 ddimensional:1 implies:1 inflated:1 direction:3 radius:3 hull:1 centered:4 require:2 fix:1 decompose:1 anonymous:1 probable:1 bregler:1 hold:1 around:3 exp:1 mapping:21 visualize:2 matthew:1 trailing:1 kambhatla:1 smallest:1 coordination:4 vice:1 correctness:1 occupying:1 trusted:5 weighted:5 gaussian:7 aim:1 avoid:1 casting:1 broader:1 ax:5 properly:1 vk:1 likelihood:3 indicates:1 contrast:2 centroid:1 am:1 dependent:1 dayan:1 entire:2 typically:1 hidden:2 relation:3 pixel:1 arg:2 orientation:1 priori:1 constrained:1 special:2 fairly:1 art:1 field:1 construct:2 equal:1 eigenproblem:1 once:1 sampling:1 represents:2 look:2 k2f:1 nearly:1 peaked:1 contaminated:3 develops:1 few:1 randomly:2 dg:1 preserve:2 tightly:1 densely:1 geometry:1 n1:1 freedom:2 ukn:1 possibility:1 guk:3 alignment:1 mixture:6 analyzed:1 yielding:2 light:1 adjoining:2 edge:2 integral:1 xy:3 respective:1 facial:1 pancake:1 penalizes:2 desired:1 minimal:2 fitted:4 merl:1 soft:3 modeling:2 asking:1 cover:3 maximization:2 cost:1 neutral:1 subset:5 uniform:1 uninteresting:1 front:1 connect:2 kn:1 jittered:1 synthetic:2 thanks:1 density:8 peak:3 off:1 uk1:1 invertible:1 connecting:3 p2k:3 together:1 intersecting:1 squared:1 containing:1 slowly:1 corner:1 return:1 potential:1 de:2 coding:1 wk:3 bold:1 satisfy:1 hasn:1 explicitly:1 root:1 view:7 lab:1 closed:7 linked:2 break:1 parallel:3 slope:2 minimize:2 chart:32 ni:1 square:3 roll:1 variance:6 efficiently:1 maximized:2 correspond:1 climbing:2 confounds:1 lkopf:1 raw:3 bayesian:1 handwritten:1 populating:1 knot:4 served:1 pcas:1 j6:5 straight:2 against:2 colleague:2 naturally:1 associated:3 mi:7 recovers:1 sampled:7 demers:2 dataset:1 treatment:1 dimensionality:16 noiselessly:1 back:5 appears:5 higher:1 isometric:2 maximally:1 arranged:1 leen:1 shrink:1 though:1 prelude:1 just:2 smola:2 spiky:1 until:1 langford:1 nonlinear:7 overlapping:1 kn2:1 defines:2 mode:2 brings:1 reveal:1 indicated:1 grows:4 usa:1 effect:1 contain:1 true:3 isomap:2 hausdorff:1 unroll:1 vicinity:1 regularization:1 iteratively:1 nonzero:2 adduced:1 attractive:1 adjacent:1 encourages:1 noted:2 criterion:3 generalized:1 gg:1 hill:2 omohundro:1 performs:1 silva:1 image:11 variational:2 wise:1 rotation:1 behaves:1 shear:1 ji:2 twist:1 volume:2 anisotropic:1 eft:2 numerically:2 synthesized:2 significant:1 cambridge:1 versa:1 smoothness:1 automatic:1 rd:7 curl:4 consistency:1 fk:2 i6:1 pointed:1 zeroing:1 grid:2 nonlinearity:1 pathology:1 had:1 dot:1 stable:1 longer:1 surface:5 add:1 dominant:3 curvature:5 posterior:11 closest:1 recent:2 isometry:1 reverse:3 nonconvex:1 inconsistency:1 yi:13 preserving:3 determine:1 maximize:1 ight:2 strike:1 signal:2 full:1 unimodal:2 smooth:4 positioning:1 faster:1 offer:1 cross:1 divided:1 gfk:1 equally:1 ensuring:1 expectation:1 kernel:4 achieved:1 folding:2 preserved:1 background:1 addition:1 grow:2 source:2 sch:1 unlike:2 finely:1 pass:2 comment:1 december:2 regularly:2 inconsistent:2 call:1 near:2 uki:2 embeddings:3 enough:1 variety:1 fit:1 gave:1 hastie:2 topology:1 inner:1 decline:1 bottleneck:1 distorting:2 expression:3 rms:1 speaking:1 constitute:1 patchwork:1 useful:2 eigenvectors:7 eigenproblems:2 amount:1 nonparametric:1 transforms:1 locally:16 tenenbaum:3 visualized:1 reduced:3 mirrored:1 estimated:2 overly:1 track:1 per:2 delta:1 broadly:1 group:1 key:1 prevent:1 gmm:9 rewriting:1 diffusion:1 backward:1 merely:1 sum:2 orient:1 angle:3 inverse:3 uncertainty:3 jitter:1 place:1 reasonable:2 patch:5 trefoil:2 dy:1 scaling:1 prefer:1 comparable:1 layer:1 fold:2 topological:1 strength:1 adapted:1 constraint:4 untied:1 flat:1 calling:1 nearby:6 dominated:1 generates:4 u1:1 innovative:1 span:3 min:2 expanded:1 px:1 conjecture:1 combination:1 ball:9 smaller:1 em:1 making:1 projecting:3 equation:11 mutually:1 count:3 fail:1 nldr:4 needed:1 xyz:1 operation:1 gaussians:1 disagreement:1 blurry:1 eigen:1 original:6 denotes:1 calculating:1 build:1 yz:1 backprojection:2 objective:2 added:2 coax:1 strategy:1 dependence:2 nr:1 diagonal:2 kth:2 subspace:10 distance:5 thank:1 outer:1 evenly:1 manifold:52 threaded:1 considers:1 unstable:1 reason:1 charting:20 code:1 length:2 relationship:2 ratio:2 unrolled:1 subproblems:1 broadway:1 trace:2 gk:8 rise:1 suppress:1 teh:2 upper:1 disagree:1 datasets:1 discarded:1 arc:2 protecting:1 january:1 immediate:1 defining:2 hinton:2 y1:2 barycentric:1 frame:4 perturbation:1 arbitrary:1 ordinate:1 nonlinearly:1 distorts:1 specified:3 connection:12 nip:4 trans:1 suggested:2 below:3 challenge:1 video:5 analogue:1 overlap:1 difficulty:1 regularized:3 indicator:1 solvable:1 buhmann:1 scheme:1 axis:2 auto:1 prior:5 geometric:2 literature:1 tangent:1 relative:2 embedded:4 loss:3 fully:1 bear:1 interesting:1 analogy:1 integrate:1 degree:3 affine:4 consistent:1 nondegenerate:2 pi:2 row:3 eccv:1 side:1 lle:15 conformally:2 neighbor:3 saul:2 sparse:1 distributed:2 curve:7 dimension:4 xn:2 boundary:4 charted:1 rich:1 computes:1 forward:4 collection:1 projected:4 avoided:2 makar:1 nonlocal:2 reconstructed:1 monotonicity:1 global:10 anchor:2 summing:1 gomes:2 xi:1 eigendecompose:1 subsequence:1 continuous:2 why:1 reasonably:1 robust:1 expanding:1 williamson:1 electric:1 diag:1 flattening:1 pk:13 dense:1 spread:1 noise:19 complementary:2 x1:1 position:1 lie:1 candidate:1 perceptual:1 down:1 removing:1 insightful:1 mojsilovic:2 virtue:1 derives:1 intrinsic:6 exists:1 quantization:1 albeit:1 effectively:3 flattened:1 magnitude:1 cartesian:1 hole:1 easier:1 locality:2 entropy:1 simply:1 forming:1 expressed:2 contained:1 tracking:1 temporarily:1 trix:1 vulnerable:1 catastrophe:1 determines:1 ma:2 goal:3 identity:1 consequently:1 revow:1 content:1 change:4 determined:1 except:1 principal:8 pas:1 catastrophic:2 brand:2 indicating:1 uneven:1 people:1 untie:1 violated:1 p21:1 regularizing:1 phenomenon:1 |
1,280 | 2,166 | Adapting Codes and Embeddings for Polychotomies
Gunnar R?atsch, Alexander J. Smola
RSISE, CSL, Machine Learning Group
The Australian National University
Canberra, 0200 ACT, Australia
Gunnar.Raetsch, Alex.Smola @anu.edu.au
Sebastian Mika
Fraunhofer FIRST
Kekulestr. 7
12489 Berlin, Germany
[email protected]
Abstract
In this paper we consider formulations of multi-class problems based on
a generalized notion of a margin and using output coding. This includes,
but is not restricted to, standard multi-class SVM formulations. Differently from many previous approaches we learn the code as well as the
embedding function. We illustrate how this can lead to a formulation
that allows for solving a wider range of problems with for instance many
classes or even ?missing classes?. To keep our optimization problems
tractable we propose an algorithm capable of solving them using twoclass classifiers, similar in spirit to Boosting.
1 Introduction
The theory of pattern recognition is primarily concerned with the case of binary classification, i.e. of assigning examples to one of two categories, such that the expected number of
misassignments is minimal. Whilst this scenario is rather well understood, theoretically as
well as empirically, it is not directly applicable to many practically relevant scenarios, the
most prominent being the case of more than two possible outcomes.
Several learning techniques naturally generalize to an arbitrary number of classes, such as
density estimation, or logistic regression. However, when comparing the reported performance of these systems with the de-facto standard of using two-class techniques in combination with simple, fixed output codes to solve multi-class problems, they often lack in
terms of performance, ease of optimization, and/or run-time behavior.
On the other hand, many methods have been proposed to apply binary classifiers to multiclass problems, such as Error Correcting Output Codes (ECOC) [6, 1], Pairwise Coupling
[9], or by simply reducing the problem of discriminating classes to ?one vs. the rest?
dichotomies. Unfortunately the optimality of such methods is not always clear (e.g., how
to choose the code, how to combine predictions, scalability to many classes).
Finally, there are other problems similar to multi-class classification which can not be
solved satisfactory by just combining simpler variants of other algorithms: multi-label
problems, where each instance should be assigned to a subset of possible categories, and
ranking problems, where each instance should be assigned a rank for all or a subset of possible outcomes. These problems can, in reverse order of their appearance, be understood as
more and more refined variants of a multi-variate regression, i.e.
two-class
multi-class
multi-label
ranking
multi-variate regression
Which framework and which algorithm in there one ever chooses, it is usually possible
to make out a single scheme common to all these: There is an encoding step in which
the input data are embedded into some ?code space? and in this space there is a code
book which allows to assign one or several labels or ranks respectively by measuring the
similarity between mapped samples and the code book entries. However, most previous
work is either focused on finding a good embedding given a fixed code or just optimizing
the code, given a fixed embedding (cf. Section 2.3).
The aim of this work is to propose (i) a multi-class formulation which optimizes the code
and the embedding of the training sample into the code space, and (ii) to develop a general
ranking technique which as well specializes to specific multi-class, multi-label and ranking
problems as it allows to solve more general problems. As an example of the latter consider
the following model problem: In chemistry people are interested in mapping sequences
to structures. It is not yet known if there is an one-to-one correspondence and hence the
problem is to find for each sequence the best matching structures. However, there are only
say a thousand sequences the chemists have good knowledge about. They are assigned,
with a certain rank, to a subset of say a thousand different structures. One could try to cast
this as a standard multi-class problem by assigning each training sequence to the structure
ranked highest. But then, there will be classes to which only very few or no sequences
are assigned and one can obviously hardly learn using traditional techniques. The machine
we propose is (at least in principle) able to solve problems like this by reflecting relations
between classes in the way the code book is constructed and at the same time trying to find
an embedding of the data space into the code space that allows for a good discrimination.
The remainder of this paper is organized as follows: In Section 2 we introduce some basic
notions of large margin, output coding and multi-class classification. Then we discuss the
approaches of [4] and [21] and propose to learn the code book. In Section 3 we propose
a rather general idea to solve resulting multi-class problems using two-class classifiers.
Section 4 presents some preliminary experiments before we conclude.
2 Large Margin Multi-Class Classification
! "
#
Denote by the sample space (not necessarily a metric space), by the space of possible
labels or ranks (e.g.
for multi-class problems where denotes the number
of classes, or
for a ranking problem), and let be a training sample of size ,
with
.
i.e.
$
Output Coding It is well known (see [6, 1] and references therein) that multi-class problems can be solved by decomposing a polychotomy into dichotomies and solving these
separately using a two-class technique. This can be understood as assigning to each class
a binary string
of length which is called a code word. This results
in an
binary code matrix. Now each of the columns of this matrix defines a partitioning of classes into two subsets, forming binary problems for which a classifier is
trained. Evaluation is done by computing the output of all learned functions, forming a
new bit-string, and then choosing the class such that some distance measure between this
string and the corresponding row of the code matrix is minimal, usually the Hamming distance. Ties can be broken by uniformly selecting a winning class, using prior information
or, where possible, using confidence outputs from the basic classifiers. 1
Since the codes for each class must be unique, there are
(for
)
possible code matrices to choose from. One possibility is to choose the codes to be errorcorrecting (ECOC) [6]. Here one uses a code book with e.g. large Hamming distance
between the code words, such that one still gets the correct decision even if a few of the
classifiers err. However, finding the code that minimizes the training error is NP-complete,
even for fixed binary classifiers [4]. Furthermore, errors committed by the binary classifiers are not necessarily independent, significantly reducing the effective number of wrong
bits that one can handle [18, 19]. Nonetheless ECOC has proven useful and algorithms
for finding a good code (and partly also finding the corresponding classifiers) have been
%
-.,
& % ' ( ) +*
,
%
,
,
1/ 0352 4 76 8 3 *
1
We could also use ternary codes, i.e.
<>=@?
ACB)AD?E , allowing for ?don?t care? classes.
9;: 8 *
proposed in e.g. [15, 7, 1, 19, 4]. Noticeably, most practical approaches suggest to drop the
requirement of binary codes, and instead propose to use continuous ones.
We now show how predictions with small (e.g. Hamming) distance to their appropriated
code words can be related to a large margin classifier, beginning with binary classification.
2.1 Large Margins
Dichotomies Here a large margin classifier is defined as a mapping
with the
property that
, or more specifically
with
, where is
some positive constant [20]. Since such a positive margin may not always be achievable,
one typically maximizes a penalized version of the maximum margin, such as
where ( ".
and (1)
! is a regularization constant and denotes the
Here is a regularization term,
( we could
( CNote
class
consideration.
that
for
the condition
(
#
(
rewrite
( (and
of" functions
( under
#
#
also
as
likewise
0
0
(
). In other words, we can express (the margin as the difference between the
from the target and the target .
for
distance of
$
& % @ %
* %
, &%(' *)
)
,+-
*
.) # !
(2)
/146
85 0327 $ & % )@ ( $ & ) #
)
This means that we
measure the minimal
relative difference in distance between , the
correct target & and any other target & % (cf. [4]). We obtain accordingly the following
Polychotomies While this insight by itself is not particularly useful, it paves the way for
an extension of the notion of the margin to multi-class problems: denote by a distance
measure and by
,
(
is the length of the code) target vectors
corresponding to class . Then we can define the margin
of an observation and
as
class with respect to
9 )
where $ & % ) #( $ & )@ # ( (3)
minimize
,
: < and ) = . For the time being we chose as a reference margin ?
for all %;
an adaptive means of choosing the reference margin can be implemented using the > -trick,
optimization problem:
which leads to an easier to control regularization parameter [16].
( ) ?0
$ &> ) -?D& @
0
Lemma 1 (Difference of Distance Measures) Denote
by
$ (& >& A ) ) )
*
*=
) a sym
)
&CB is convex in for all
metric distance measure.
Then the only case where $ >&
& &6B occurs if $ &> ) <$D & $E *) &GFIH ) , where H
*KJ * is symmetric.
) ( $ &MB1 ) is positive semidefinite. This is
)
Proof Convexity in implies that L E 0 $ &>
) is a function of ) only. The latter, however, implies that the only
only possible if L E 0 $ ) &>
)
joint terms in & and must be of linear nature in . Symmetry, on the other hand, implies
2.2 Distance Measures
Several choices of are possible. However, one can show that only
and related functions will lead to a convex constraint on :
$
)
&
that the term must be linear in , too, which proves the claim.
Lemma 1 implies that any distance functions other than the ones described above will lead
to optimization problems with potentially many local minima, which is not desirable. However, for quadratic we will get a convex optimization problem (assuming suitable
)
$
& % & % H means
$ &> ) ?D& ( ) ? 0
$ & % )@ ( $ & )@ ?D& % ? 0 ( ?D& ? 0 8 & % F )@ ( 8 & F )@ (4)
Note,
if the code words have the same length, the difference of the projections of
) ontothat
different code
We will indeed later consider a more
words
)@ determines
& % F ) the margin.
convenient case: $ & %
, which will lead to linear constraints only and
and then there are ways to efficiently solve (3). Finally, re-defining
that it is sufficient to consider only
. We obtain
allows us to use standard optimization packages. However, there are no principal limitations
about using the Euclidean distance.
If we choose
to be an error-correcting code, such as those in [6, 1], one will often have
. Hence, we use fewer dimensions than we have classes. This means that during
optimization we are trying to find
functions
,
,
from an dimensional subspace. In other words, we choose the subspace and perform
regularization by allowing only a smaller class of functions. By appropriately choosing the
subspace one may encode prior knowledge about the problem.
,
& %
4 $ & % ) # % )
,
( 8
8 7 ( 4 & % 4
8 & % F )@ ( 8 & F )@
2.3 Discussion and Relation to Previous Approaches
we have that (4) is equal
Note that for
and hence the problem of multi-class classification reverts to the problem
of solving binary classification problems of one vs. the remaining classes. Then our approach turns out to be very similar to the idea presented in [21] (except for some additional
slack-variables).
A different approach was taken in [4]. Here, the function is held fix and the code is
optimized. In their approach, the code is described as a vector in a kernel feature space and
one obtains in fact an optimization problem very similar to the one in [21] and (3) (again,
the slack-variables are defined slightly different).
Another idea which is quite similar to ours was also presented at the conference [5]. The
resulting optimization problem turns out to be convex, but with the drawback, that one can
either not fully optimize the code vectors or not guarantee that they are well separated.
Since these approaches were motivated by different ideas (one optimizing the code, the
other optimizing the embedding), this shows that the role of the code
and the embedding function is interchangeable if the function or the code, respectively, is fixed.
Our approach allows arbitrary codes for which a function is learned. This is illustrated in
Figure 1. The position of the code words (=?class centers?) determine the function . The
position of the centers relative to each other may reflect relationships between the classes
(e.g. classes ?black? & ?white? and ?white? & ?grey? are close).
)
)
)
&
& %
)
)
Figure 1: Illustration of embedding idea: The samples are mapped from the input space
into the code space via the embedding function , such that samples from the same class
are close to their respective code book vector (crosses on the right). The spatial organization
of the code book vectors reflects the organization of classes in the space.
)
2.4 Learning Code & Embedding
This leaves us with the question of how to determine a ?good? code and a suitable . As we
can see from (4), for fixed the constraints are linear in and vice versa, yet we have non-
)
&
)
&
)
&
convex constraints, if both and are variable. Finding the global optimum is therefore
computationally infeasible when optimizing and simultaneously (furthermore note that
any rotation applied to and will leave the margin invariant, which shows the presence
of local minima due to equivalent codes).
Instead, we propose the following method: for fixed code optimize over , and subsequently, for fixed , optimize over , possibly repeating the process. The first step follows
[4], i.e. to learn the code for a fixed function. Both steps separately can be performed fairly
efficient (since the optimization problems are convex; cf. Lemma 1).
This procedure is guaranteed to decrease the over all objective function at every step and
converges to a local minimum. We now show how a code maximizing the margin can be
found. To avoid a trivial solution (we can may virtually increase the margin by rescaling all
by some constant), we add
to the objective function. It can be shown that
one does not need an additional regularization constant in front of this term, if the distance
is linear on both arguments. If one prefers sparse codes, one may use the -norm instead.
In summary, we obtain the following convex quadratic program for finding the codes which
can be solved using standard optimization techniques:
&
)
)
)
&
&
4
?D& % ? 00
&
D 4
minimize
subject to
%
( 4#
) ?D& ?00 (
& & % F
"
and % :
for all .
(5)
The technique for finding the embedding will be discussed in more detail in Section 3.
Initialization To obtain a good initial code, we may either take recourse to readily available tables [17] or we may use a random code, e.g. by generating vectors uniformly
distributed on the -dimensional sphere. One can show that the probability that there
exists two such vectors (out of ) that have a smaller distance than is bounded by
(proof given in the full paper). Hence, with probability greater
than the random code vectors have distances greater than
from each other.2
(
0 9 9
0
,
0 2 * 8)
8
3 Column Generation for Finding the Embedding
There are several ways to setup and optimize the resulting optimization problem (3). For
instance in [21, 4] the class of functions is the set of hyperplanes in some kernel feature
space and the regularizer
is the sum of the -norms of the hyperplane
normal vectors.
!"
$# %
In this section we consider a different approach. Denote by
)(
" -, "./"
+*
a class of basis functions and let
'&
'01
2 . We choose
the regularizer
to be the -norm on the expansion coefficients. We are interested in
solving:
)
0
)
2
+
* , "
/ 03203/10435 "
687!9-:; <.7!9>=
?
(6)
(
(
& % & F 6
? ".
# : % )
subject to
To derive a column generation method
the dual optimization problem, or
A 4 @[12,
, " 2] we)
need
# and : % )
,
more specifically its constraints:
A 4 & ( & % # F A" CB )-% )
&
(7)
7 @
5 4
2
However, also note that this is quite a bit worse than the best packing, which scales with DCE
rather than D E . This is due a the union-bound argument in the proof, which requires us to sum
IKJ.L pairs have more than M distance.
over the probability that all DGFHD
.= ?
7
5 4
A 4 B
" )
#
and
. The idea of column generation is to start with a
,
restricted master problem, namely without the variables (i.e &
). Then one solves the
corresponding dual problem (7) and then finds the hypothesis that corresponds to a violated
constraint (and also one primal variable). This hypothesis is included in the optimization
problem, one resolves and finds the next violated constraint. If all constraints of the full
problem are satisfied, one has reached optimality.
"@
We now construct a hypothesis set from a scalar valued base-class
& , which has particularly nice properties for our purposes. The idea
where %
"
is to extend @ by multiplication with vectors
:
+
*
*
* ? ? )
+
Since there are infinitely many functions in this set , we have an infinite number of constraints in the dual optimization problem. By using the described column generation technique one can, however, find the solution of this semi-infinite programming problem [13].
We have to identify the constraint in" (7), which is maximally violated, i.e. one has to find a
?partitioning? and a hypothesis @ with maximal
A 4 "@ & ( & % F
F "@
(8)
7
5 4
"
"
for appropriate @ . Maximizing (8) with respect to ? ? is easy for a given @ :
8
"@
"@
"@
for , one chooses
2 ; if , then ? ? 0 and for
one chooses the minimizing" unit vector. However, finding and simultaneously is a
difficult problem, if not all @ are known in advance (see also [15]). We propose to test all
"
previously used hypotheses to find the best . As a second step one finds the hypothesis @
"@
that maximizes
. Only if one cannot find a hypothesis that violates a constraint, one
employs the more sophisticated techniques suggested in [15]. If there is no hypothesis
left that corresponds to a violated constraint, the dual optimization problem is optimal.
In this work we are mainly
interested in the case
, since then
and the
"
problem of finding @ simplifies greatly. Then we can use another learning algorithm that
minimizes or approximately minimizes the training error of a weighted training set (rewrite
(8)). This approach has indeed many similarities to Boosting. Following the ideas in [14]
one can show that there is a close relationship between our technique using the trivial code
and the multi-class boosting algorithms as e.g. proposed in [15].
F
>*
4 Extensions and Illustration
4.1 A first Experiment
In a preliminary set of experiments we use two benchmark data sets from the UCI benchmark repository: glass and iris. We used our column generation strategy as described
in Section 3 in conjunction with the code optimization problem to solve the combined optimization problem to find the code and the embedding. We used
. The algorithm has
only one model parameters ( ). We selected it by -fold cross validation on the training
data. The test error is determined by averaging over five splits of training and test data.
As base learning algorithm we chose decision trees (C4.5) which we only use as two-class
classifier in our column generation algorithm.
On the glass data set we obtained an error rate of
. In [1] an error of
was reported for SVMs using a polynomial kernel. We also computed the test error of multiclass decision trees and obtained
error. Hence, our hybrid algorithm could
relatively improve existing results by
. On the iris data we could achieve an error
rate of
and could slightly improve the result of decision trees (
).
" 8 #
!
,
8
However, SVMs beat our result with
error [1]. We conjecture that this is due to the
properties of decision trees which have problems generating smooth boundaries not aligned
with coordinate axes.
So far, we could only show a proof of concept and more experimental work is necessary.
It is in particular interesting to find practical examples, where a non-trivial choice of the
code (via optimization) helps simplifying the embedding and finally leads to additional
improvements. Such problems often appear in Computer Vision, where there are strong relationships between classes. Preliminary results indicate that one can achieve considerable
improvements when adapting codes and embeddings [3].
3
Figure 2: Toy example for learning missing classes. Shown is
the decision boundary and the confidence for assigning a sample
to the upper left class. The training set, however, did not contain
samples from this class. Instead, we used (9) with the information that each example besides belonging to its own class with
confidence two also belongs to the other classes with confidence
one iff its distance to the respective center is less than one.
2.5
2
1.5
1
0.5
0.5
1
1.5
2
2.5
3
4.2 Beyond Multi-Class
So far we only considered the case where there is only one class to which an example
belongs to. In a more general setting as for example the problem mentioned in the introduction, there can be several classes, which possibly have a ranking. We have the sets
, which contain all pairs of ?relations? between
#
contains all pairs of positive and negative classes of
the positive classes. The set
an example.
% % 0 % % 0
D#
* , "
% ? 0
B
?
&
minimize
?
?
"
0
B
' A
5& A %
& % ( & % # F ) 6 (
(9)
subject to
? %
0
%
& % for
( all& % ".# F ) 6
and
( B 0
?
0
for all ".
and % %
) "*
, ".A" and 7 !A" +
*#% )
0 & .
where 6
)
&
In this formulation one tries to find a code and an embedding , such that for each example the output wrt. each class this example has a relation with, reflects the order of this
relations (i.e. the examples get ranked appropriately). Furthermore, the program tries to
achieve a ?large margin? between relevant and irrelevant classes for each sample. Similar
formulations can be found in [8] (see also [11]).
Optimization of (9) is analogous to the column generation approach discussed in Section 3.
We omit details due to constraints on space. A small toy example, again as a limited proof
of concept, is given in Figure 2.
Connection to Ranking Techniques Ordinal regression through large margins [10] can
be seen as an extreme case of (9), where we have as many classes as" observations,
and each
"
pair of observations
has
to
satisfy
a
ranking
relation
,
if
is to
"
be preferred to . This formulation can of course also be understood as a special case of
multi-dimensional regression.
( (
5 Conclusion
We proposed an algorithm to simultaneously optimize output codes and the embedding of
the sample into the code book space building upon the notion of large margins. Further-
more, we have shown, that only quadratic and related distance measures in the code book
space will lead to convex constraints and hence convex optimization problems whenever
either the code or the embedding is held fixed. This is desirable since at least for these
sub-problems there exist fairly efficient techniques to compute these (of course the combined optimization problem of finding the code and the embedding is not convex and has
local minima). We proposed a column generation technique for solving the embedding optimization problems. It allows the use of a two-class algorithm, of which there exists many
efficient ones, and has connection to boosting. Finally we proposed a technique along the
same lines that should be favorable when dealing with many classes or even empty classes.
Future work will concentrate on finding more efficient algorithms to solve the optimization
problem and on more carefully evaluating their performance.
Acknowledgements We thank B. Williamson and A. Torda for interesting discussions.
References
[1] E.L. Allwein, R.E. Schapire, and Y. Singer. Reducing multiclass to binary: A unifying approach
for margin classifiers. Journal of Machine Learning Research, 1:113?141, 2000.
[2] K.P. Bennett, A. Demiriz, and J. Shawe-Taylor. A column generation algorithm for boosting.
In P. Langley, editor, Proc. 17th ICML, pages 65?72, San Francisco, 2000. Morgan Kaufmann.
[3] B. Caputo and G. R?atsch. Adaptive codes for visual categories. November 2002. Unpublished
manuscript. Partial results presented at NIPS?02.
[4] K. Crammer and Y. Singer. On the learnability and design of output codes for multiclass problems. In N. Cesa-Bianchi and S. Goldberg, editors, Proc. Colt, pages 35?46, San Francisco,
2000. Morgan Kaufmann.
[5] O. Dekel and Y. Singer. Multiclass learning by probabilistic embeddings. In NIPS, vol. 15. MIT
Press, 2003.
[6] T.G. Dietterich and G. Bakiri. Solving multiclass learning problems via error-correcting output
codes. Journal of Aritifical Intelligence Research, 2:263?286, 1995.
[7] V. Guruswami and A. Sahai. Multiclass learning, boosing, and error-correcting codes. In
Proc. of the twelfth annual conference on Computational learning theory, pages 145?155, New
York, USA, 1999. ACM Press.
[8] S. Har-Peled, D. Roth, and D. Zimak. Constraint classification: A new approach to multiclass
classification and ranking. In NIPS, vol. 15. MIT Press, 2003.
[9] T.J. Hastie and R.J. Tibshirani. Classification by pairwise coupling. In M.I. Jordan, M.J.
Kearnsa, and S.A. Solla, editors, Advances in Neural Information Processing Systems, vol. 10.
MIT Press, 1998.
[10] R. Herbrich, T. Graepel, and K. Obermayer. Large margin rank boundaries for ordinal regression. In A. J. Smola, P. L. Bartlett, B. Sch?olkopf, and D. Schuurmans, editors, Advances in
Large Margin Classifiers, pages 115?132, Cambridge, MA, 2000. MIT Press.
[11] R. Jin and Z. Ghahramani. Learning with multiple labels. In NIPS, vol. 15. MIT Press, 2003.
[12] S. Nash and A. Sofer. Linear and Nonlinear Programming. McGraw-Hill, New York, NY,
1996.
[13] G. R?atsch, A. Demiriz, and K. Bennett. Sparse regression ensembles in infinite and finite
hypothesis spaces. Machine Learning, 48(1-3):193?221, 2002. Special Issue on New Methods
for Model Selection and Model Combination.
[14] G. R?atsch, M. Warmuth, S. Mika, T. Onoda, S. Lemm, and K.-R. M?uller. Barrier boosting. In
Proc. COLT, pages 170?179, San Francisco, 2000. Morgan Kaufmann.
[15] R.E. Schapire. Using output codes to boost multiclass learning problems. In Machine Learning:
Proceedings of the 14th International Conference, pages 313?321, 1997.
[16] B. Sch?olkopf, A. Smola, R.C. Williamson, and P.L. Bartlett. New support vector algorithms.
Neural Computation, 12:1207 ? 1245, 2000.
[17] N. Sloane. Personal homepage. http://www.research.att.com/?njas/.
[18] W. Utschick. Error-Correcting Classification Based on Neural Networks. Shaker, 1998.
[19] W. Utschick and W. Weichselberger. Stochastic organization of output codes in multiclass
learning problems. Neural Computation, 13(5):1065?1102, 2001.
[20] V.N. Vapnik and A.Y. Chervonenkis. A note on one class of perceptrons. Automation and
Remote Control, 25, 1964.
[21] J. Weston and C. Watkins. Multi-class support vector machines. Technical Report CSD-TR-9804, Royal Holloway, University of London, Egham, 1998.
| 2166 |@word repository:1 version:1 achievable:1 norm:3 mb1:1 polynomial:1 dekel:1 twelfth:1 grey:1 gfih:1 simplifying:1 tr:1 initial:1 contains:1 att:1 selecting:1 chervonenkis:1 ours:1 existing:1 err:1 comparing:1 com:1 assigning:4 yet:2 must:3 readily:1 drop:1 v:2 discrimination:1 intelligence:1 fewer:1 leaf:1 selected:1 warmuth:1 accordingly:1 beginning:1 boosting:6 herbrich:1 hyperplanes:1 simpler:1 five:1 along:1 constructed:1 combine:1 introduce:1 pairwise:2 theoretically:1 weichselberger:1 indeed:2 expected:1 behavior:1 multi:24 ecoc:3 resolve:1 csl:1 bounded:1 maximizes:2 homepage:1 polychotomies:2 string:3 minimizes:3 whilst:1 finding:12 guarantee:1 every:1 act:1 tie:1 classifier:14 wrong:1 facto:1 partitioning:2 control:2 unit:1 omit:1 appear:1 before:1 positive:5 understood:4 local:4 encoding:1 approximately:1 mika:3 chose:2 black:1 au:1 therein:1 initialization:1 ease:1 limited:1 range:1 unique:1 practical:2 ternary:1 union:1 procedure:1 langley:1 adapting:2 significantly:1 matching:1 projection:1 word:8 confidence:4 convenient:1 suggest:1 get:3 cannot:1 close:3 selection:1 optimize:5 equivalent:1 www:1 missing:2 center:3 maximizing:2 roth:1 convex:10 focused:1 correcting:5 insight:1 embedding:19 handle:1 notion:4 coordinate:1 sahai:1 analogous:1 target:5 programming:2 us:1 goldberg:1 hypothesis:9 trick:1 recognition:1 particularly:2 role:1 solved:3 thousand:2 solla:1 decrease:1 highest:1 remote:1 mentioned:1 broken:1 convexity:1 peled:1 nash:1 chemist:1 personal:1 trained:1 solving:7 rewrite:2 interchangeable:1 upon:1 basis:1 packing:1 joint:1 differently:1 regularizer:2 separated:1 effective:1 london:1 dichotomy:3 outcome:2 refined:1 choosing:3 quite:2 solve:7 valued:1 say:2 demiriz:2 itself:1 obviously:1 sequence:5 propose:8 maximal:1 remainder:1 relevant:2 combining:1 uci:1 aligned:1 iff:1 achieve:3 scalability:1 olkopf:2 empty:1 requirement:1 optimum:1 generating:2 leave:1 converges:1 wider:1 illustrate:1 coupling:2 develop:1 derive:1 help:1 strong:1 solves:1 implemented:1 implies:4 australian:1 indicate:1 concentrate:1 drawback:1 correct:2 subsequently:1 stochastic:1 australia:1 violates:1 noticeably:1 assign:1 fix:1 preliminary:3 extension:2 practically:1 considered:1 normal:1 cb:2 mapping:2 claim:1 purpose:1 estimation:1 favorable:1 proc:4 applicable:1 label:6 vice:1 reflects:2 weighted:1 uller:1 mit:5 always:2 aim:1 rather:3 avoid:1 allwein:1 conjunction:1 encode:1 ax:1 improvement:2 rank:5 mainly:1 greatly:1 glass:2 typically:1 relation:6 fhg:1 interested:3 germany:1 issue:1 classification:11 dual:4 colt:2 spatial:1 special:2 fairly:2 equal:1 construct:1 icml:1 future:1 np:1 report:1 employ:1 primarily:1 few:2 simultaneously:3 national:1 organization:3 possibility:1 evaluation:1 extreme:1 semidefinite:1 primal:1 held:2 har:1 capable:1 partial:1 necessary:1 respective:2 tree:4 euclidean:1 taylor:1 re:1 minimal:3 instance:4 column:10 zimak:1 measuring:1 subset:4 entry:1 too:1 front:1 learnability:1 reported:2 acb:1 chooses:3 combined:2 density:1 international:1 discriminating:1 probabilistic:1 sofer:1 again:2 reflect:1 satisfied:1 cesa:1 choose:6 possibly:2 worse:1 book:9 rescaling:1 toy:2 de:2 chemistry:1 coding:3 includes:1 coefficient:1 automation:1 satisfy:1 ranking:9 ad:1 later:1 try:3 performed:1 reached:1 start:1 minimize:3 kekulestr:1 kaufmann:3 likewise:1 efficiently:1 ensemble:1 identify:1 generalize:1 sebastian:1 whenever:1 nonetheless:1 naturally:1 proof:5 hamming:3 knowledge:2 dce:1 organized:1 graepel:1 sophisticated:1 carefully:1 reflecting:1 manuscript:1 maximally:1 formulation:7 done:1 furthermore:3 just:2 smola:4 hand:2 nonlinear:1 lack:1 defines:1 logistic:1 building:1 dietterich:1 usa:1 concept:2 contain:2 hence:6 assigned:4 regularization:5 symmetric:1 satisfactory:1 illustrated:1 white:2 during:1 iris:2 generalized:1 prominent:1 trying:2 hill:1 sloane:1 complete:1 consideration:1 common:1 rotation:1 empirically:1 discussed:2 extend:1 raetsch:1 versa:1 cambridge:1 shawe:1 similarity:2 add:1 base:2 own:1 optimizing:4 optimizes:1 belongs:2 reverse:1 scenario:2 irrelevant:1 certain:1 binary:11 seen:1 minimum:4 additional:3 care:1 greater:2 morgan:3 utschick:2 determine:2 ii:1 semi:1 full:2 desirable:2 multiple:1 smooth:1 technical:1 cross:2 sphere:1 prediction:2 variant:2 regression:7 basic:2 vision:1 metric:2 kernel:3 separately:2 appropriately:2 sch:2 rest:1 subject:3 virtually:1 spirit:1 jordan:1 presence:1 split:1 embeddings:3 concerned:1 easy:1 variate:2 hastie:1 idea:8 simplifies:1 twoclass:1 multiclass:10 motivated:1 guruswami:1 bartlett:2 york:2 hardly:1 prefers:1 useful:2 clear:1 repeating:1 svms:2 category:3 schapire:2 http:1 exist:1 tibshirani:1 vol:4 express:1 group:1 gunnar:2 sum:2 run:1 package:1 master:1 decision:6 bit:3 bound:1 guaranteed:1 correspondence:1 fold:1 quadratic:3 annual:1 constraint:15 alex:1 lemm:1 argument:2 optimality:2 relatively:1 conjecture:1 combination:2 belonging:1 smaller:2 slightly:2 ikj:1 restricted:2 invariant:1 errorcorrecting:1 taken:1 recourse:1 computationally:1 previously:1 discus:1 turn:2 slack:2 wrt:1 ordinal:2 singer:3 tractable:1 available:1 decomposing:1 apply:1 appropriate:1 egham:1 denotes:2 remaining:1 cf:3 unifying:1 ghahramani:1 prof:1 bakiri:1 objective:2 question:1 occurs:1 strategy:1 traditional:1 pave:1 obermayer:1 subspace:3 distance:18 thank:1 mapped:2 berlin:1 trivial:3 assuming:1 code:71 length:3 besides:1 relationship:3 illustration:2 minimizing:1 setup:1 unfortunately:1 difficult:1 potentially:1 negative:1 design:1 perform:1 allowing:2 upper:1 bianchi:1 observation:3 benchmark:2 finite:1 november:1 jin:1 beat:1 defining:1 ever:1 committed:1 arbitrary:2 cast:1 rsise:1 pair:4 namely:1 optimized:1 connection:2 unpublished:1 c4:1 learned:2 boost:1 nip:4 able:1 suggested:1 beyond:1 usually:2 pattern:1 program:2 reverts:1 royal:1 suitable:2 ranked:2 hybrid:1 scheme:1 improve:2 specializes:1 fraunhofer:1 kj:1 prior:2 nice:1 acknowledgement:1 multiplication:1 relative:2 embedded:1 fully:1 generation:9 limitation:1 interesting:2 proven:1 validation:1 sufficient:1 principle:1 editor:4 row:1 course:2 penalized:1 summary:1 sym:1 infeasible:1 barrier:1 sparse:2 distributed:1 boundary:3 dimension:1 evaluating:1 adaptive:2 san:3 far:2 obtains:1 preferred:1 mcgraw:1 keep:1 dealing:1 global:1 conclude:1 francisco:3 don:1 continuous:1 table:1 learn:4 nature:1 onoda:1 symmetry:1 caputo:1 schuurmans:1 expansion:1 williamson:2 necessarily:2 did:1 csd:1 canberra:1 ny:1 sub:1 position:2 winning:1 watkins:1 specific:1 svm:1 exists:2 vapnik:1 anu:1 margin:23 easier:1 aritifical:1 simply:1 appearance:1 infinitely:1 forming:2 visual:1 scalar:1 corresponds:2 determines:1 acm:1 ma:1 weston:1 bennett:2 considerable:1 included:1 specifically:2 except:1 reducing:3 uniformly:2 hyperplane:1 infinite:3 determined:1 lemma:3 principal:1 called:1 averaging:1 partly:1 experimental:1 atsch:4 perceptrons:1 holloway:1 people:1 support:2 latter:2 crammer:1 alexander:1 violated:4 |
1,281 | 2,167 | A Model for Learning Variance Components of
Natural Images
Michael S. Lewicki?
[email protected]
Yan Karklin
[email protected]
Computer Science Department &
Center for the Neural Basis of Cognition
Carnegie Mellon University
Abstract
We present a hierarchical Bayesian model for learning efficient codes of
higher-order structure in natural images. The model, a non-linear generalization of independent component analysis, replaces the standard assumption of independence for the joint distribution of coefficients with
a distribution that is adapted to the variance structure of the coefficients
of an efficient image basis. This offers a novel description of higherorder image structure and provides a way to learn coarse-coded, sparsedistributed representations of abstract image properties such as object
location, scale, and texture.
1 Introduction
One of the major challenges in vision is how to derive from the retinal representation
higher-order representations that describe properties of surfaces, objects, and scenes. Physiological studies of the visual system have characterized a wide range of response properties, beginning with, for example, simple cells and complex cells. These, however, offer
only limited insight into how higher-order properties of images might be represented or
even what the higher-order properties might be. Computational approaches to vision often derive algorithms by inverse graphics, i.e. by inverting models of the physics of light
propagation and surface reflectance properties to recover object and scene properties. A
drawback of this approach is that, because of the complexity of modeling, only the simplest and most approximate models are computationally feasible to invert and these often
break down for realistic images. A more fundamental limitation, however, is that this formulation of the problem does not explain the adaptive nature of the visual system or how it
can learn highly abstract and general representations of objects and surfaces.
An alternative approach is to derive representations from the statistics of the images themselves. This information theoretic view, called efficient coding, starts with the observation
that there is an equivalence between the degree of structure represented and the efficiency
of the code [1]. The hypothesis is that the primary goal of early sensory coding is to encode information efficiently. This theory has been applied to derive efficient codes for
? To
whom correspondence should be addressed
natural images and to explain a wide range of response properties of neurons in the visual
cortex [2?7].
Most algorithms for learning efficient representations assume either simply that the data
are generated by a linear superposition of basis functions, as in independent component
analysis (ICA), or, as in sparse coding, that the basis function coefficients are ?sparsified?
by lateral inhibition. Clearly, these simple models are insufficient to capture the rich structure of natural images, and although they capture higher-order statistics of natural images
(correlations beyond second order), it remains unclear how to go beyond this to discover
higher-order image structure.
One approach is to learn image classes by embedding the statistical density assumed by
ICA in a mixture model [8]. This provides a method for modeling classes of images and
for performing automatic scene segmentation, but it assumes a fundamentally local representation and therefore is not suitable for compactly describing the large degree of structure
variation across images. Another approach is to construct a specific model of non-linear
features, e.g. the responses of complex cells, and learn an efficient code of their outputs [9].
With this, one is limited by the choice of the non-linearity and the range of image regularities that can be modeled.
In this paper, we take as a starting point the observation by Schwartz and Simoncelli [10]
that, for natural images, there are significant statistical dependencies among the variances
of filter outputs. By factoring out these dependencies with divisive normalization, Schwartz
and Simoncelli showed that the model could account for a wide range of non-linearities
observed in neurons in the auditory nerve and primary visual cortex.
Here, we propose a statistical model for higher-order structure that learns a basis on the
variance regularities in natural images. This higher-order, non-orthogonal basis describes
how, for a particular visual image patch, image basis function coefficient variances deviate
from the default assumption of independence. This view offers a novel description of
higher-order image structure and provides a way to learn sparse distributed representations
of abstract image properties such as object location, scale, and surface texture.
Efficient coding of natural images
The computational goal of efficient coding is to derive from the statistics of the pattern
ensemble a compact code that maximally reduces the redundancy in the patterns with minimal loss of information. The standard model assumes that the data is generated using a set
of basis functions A and coefficients u:
x = Au ,
(1)
Because coding efficiency is being optimized, it is necessary, either implicitly or explicitly,
for the model to capture the probability distribution of the pattern ensemble. For the linear
model, the data likelihood is [11, 12]
p(x|A) = p(u)/| det A| .
(2)
The coefficients ui , are assumed to be statistically independent
p(u) = ? p(ui ) .
(3)
i
ICA learns efficient codes of natural scenes by adapting the basis vectors to maximize
the likelihood of the ensemble of image patterns, p(x1 , . . . , xN ) = ?n p(xn |A), which maximizes the independence of the coefficients and optimizes coding efficiency within the limits
of the linear model.
a
b
c
Figure 1: Statistical dependencies among natural image independent component basis coefficients. The scatter plots show for the two basis functions in the same row and column
the joint distributions of basis function coefficients. Each point represents the encoding of
a 20 ? 20 image patch centered at random locations in the image. (a) For complex natural
scenes, the joint distributions appear to be independent, because the joint distribution can
be approximated by the product of the marginals. (b) Closer inspection of particular image
regions (the image in (b) is contained in the lower middle part of the image in (a)) reveals
complex statistical dependencies for the same set of basis functions. (c) Images such as
texture can also show complex statistical dependencies.
Statistical dependencies among ?independent? components
A linear model can only achieve limited statistical independence among the basis function
coefficients and thus can only capture a limited degree of visual structure. Deviations from
independence among the coefficients reflect particular kinds of visual structure (fig. 1). If
the coefficients were independent it would be possible to describe the joint distribution as
the product of two marginal densities, p(ui , u j ) = p(ui )p(u j ). This is approximately true
for natural scenes (fig.1a), but for particular images, the joint distribution of coefficients
show complex statistical dependencies that reflect the higher-order structure (figs.1b and
1c). The challenge for developing more general models of efficient coding is formulating
a description of these higher-order correlations in a way that captures meaningful higherorder visual structure.
2 Modeling higher-order statistical structure
The basic model of standard efficient coding methods has two major limitations. First,
the transformation from the pattern to the coefficients is linear, so only a limited class
of computations can be achieved. Second, the model can capture statistical relationships
among the pixels, but does not provide any means to capture higher order relationships
that cannot be simply described at the pixel level. As a first step toward overcoming these
limitations, we extend the basic model by introducing a non-independent prior to model
higher-order statistical relationships among the basis function coefficients.
Given a representation of natural images in terms of a Gabor-wavelet-like representation
learned by ICA, one salient statistical regularity is the covariation of basis function coefficients in different visual contexts. Any specific type of image region, e.g. a particular kind
of texture, will tend to yield in large values for some coefficients and not others. Different
types of image regions will exhibit different statistical regularities among the variances of
the coefficients. For a large ensemble of images, the goal is to find a code that describes
these higher-order correlations efficiently.
In the standard efficient coding model, the coefficients are often assumed to follow a generalized Gaussian distribution
q
p(ui ) = ze?|ui /?i | ,
(4)
where z = q/(2?i ?[1/q]). The exponent q determines the distribution?s shape and weight
of the tails, and can be fixed or estimated from the data for each basis function coefficient.
The parameter ?i determines the scale of variation (usually fixed in linear models, since
basis vectors in A can absorb the scaling). ?i is a generalized notion of variance; for clarity,
we refer to it simply as variance below.
Because we want to capture regularities among the variance patterns of the coefficients,
we do not want to model the values of u themselves. Instead, we assume that the relative
variances in different visual contexts can be modeled with a linear basis as follows
?i = exp([Bv]i )
(5)
? = Bv .
? log?
(6)
where [Bv]i refers to the ith element of the product vector Bv. This formulation is useful
because it uses a basis to represent the deviation from the variance assumed by the standard model. If we assume that vi also follows a zero-centered, sparse distribution (e.g. a
generalized Gaussian), then Bv is peaked around zero which yields a variance of one, as
in standard ICA. Because the distribution is sparse, only a few of the basis vectors in B
are needed to describe how any particular image deviates from the default assumption of
independence. The joint distribution for the prior (eqn.3) becomes
L
ui q
? log p(u|B, v) ? ? [Bv] ,
(7)
e i
i
Having formulated the problem as a statistical model, the choice of the value of v for a
given u is determined by maximizing the posterior distribution
v? = argmax p(v|u, B) = argmax p(u|B, v)p(v)
(8)
v
v
Unfortunately, computing the most probable v is not straightforward. Because v specifies
the variance of u, there is a range of values that could account for a given pattern ? all that
changes is the probability of the first order representation, p(u|B, v). For the simulations
below, v? was estimated by gradient ascent.
By maximizing the posterior p(v|u, B), the algorithm is computing the best way to describe
how the distribution of vi ?s for the current image patch deviates from the default assumption
of independence, i.e. v = 0. This aspect of the algorithm makes the transformation from
the data to the internal representation fundamentally non-linear. The basis functions in
B represent an efficient, sparse, distributed code for commonly observed deviations. In
contrast to the first layer, where basis functions in A correspond to specific visual features,
higher-order basis functions in B describe the shapes of image distributions.
The parameters are adapted by performing gradient ascent on the data likelihood. Using the
generalized prior, the data likelihood is computed by marginalizing over the coefficients.
Assuming independence between B and v, the marginal likelihood is
p(x|A, B) =
Z
p(u|B, v)p(v)/| det A|dv .
(9)
This, however, is intractable to compute, so we approximate it by the maximum a posteriori
value v?
p(x|A, B) ? p(u|B, v? )p(?v)/| detA| .
(10)
We assume that p(v) = ?i p(vi ) and that p(vi ) ? exp(?|vi |). We adapt B by maximizing
the likelihood over the data ensemble
B = argmax ? log p(un |B, v? n ) + log p(B)
(11)
B
n
For reasons of space, we omit the (straightforward) derivations of the gradients.
Figure 2: A subset of the 400 image basis functions. Each basis function is 20x20 pixels.
3 Results
The algorithm described above was applied to a standard set of ten 512?512 natural images
used in [2]. For computational simplicity, prior to the adaptation of the higher-order basis
B, a 20 ? 20 ICA image basis was derived using standard methods (e.g. [3]). A subset of
these basis functions is shown in fig. 2.
Because of the computational complexity of the learning procedure, the number of basis
functions in B was limited to 30, although in principle a complete basis of 400 could be
learned. The basis B was initialized to small random values and gradient ascent was performed for 4000 iterations, with a fixed step size of 0.05. For each batch of 5000 randomly
sampled image patches, v? was derived using 50 steps of gradient ascent at a fixed step size
of 0.01.
Fig. 3 shows three different representations of the basis functions in the matrix B adapted to
natural images. The first 10 ? 3 block (fig.3a) shows the values of the 30 basis functions in
B in their original learned order. Each square represents 400 weights B i, j from a particular
v j to all the image basis functions ui ?s. Black dots represent negative weights; white,
positive weights. In this representation, the weights appear sparse, but otherwise show no
apparent structure, simply because basis functions in A are unordered.
Figs. 3b and 3c show the weights rearranged in two different ways. In fig. 3b, the dots representing the same weights are arranged according to the spatial location within an image
patch (as determined by fitting a 2D Gabor function) of the basis function which the weight
affects. Each weight is shown as a dot; white dots represent positive weights, black dots
negative weights. In fig. 3c, the same weights are arranged according to the orientation and
spatial scale of the Gaussian envelope of the fitted Gabor. Orientation ranges from 0 to ?
counter-clockwise from the horizontal axis, and spatial scale ranges radially from DC at the
bottom center to Nyquist. (Note that the learned basis functions can only be approximately
fit by Gabor functions, which limits the precision of the visualizations.)
In these arrangements, several types of higher-order regularities emerge. The predominant
one is that coefficient variances are spatially correlated, which reflects the fact that a common occurrence is an image patch with a small localized object against a relatively uniform
background. For example, the pattern in row 5, column 3 of fig. 3b shows that often the
coefficient variances in the top and bottom halves of the image patch are anti-correlated,
i.e. either the object or scene is primarily across the top or across the bottom. Because
vi can be positive or negative, the higher-order basis functions in B represent contrast in
the variance patterns. Other common regularities are variance-contrasts between two orientations for all spatial positions (e.g. row 7, column 1) and between low and high spatial
scales for all positions and orientations (e.g. row 9, column 3). Most higher-order basis
functions have simple structure in either position, orientation, or scale, but there are some
whose organization is less obvious.
a
b
c
Figure 3: The learned higher-order basis functions. The same weights shown in the original
order (a); rearranged according to the spatial location of the corresponding image basis
functions (b); rearranged according to frequency and orientation of image basis functions
(c). See text for details.
Figure 4: Image patches that yielded the largest coefficients for two basis functions in B.
The central block contains nine image patches corresponding to higher-order basis function
coefficients with values near zero, i.e. small deviations from independent variance patterns.
Positions of other nine-patch blocks correspond to the associated values of higher-order
coefficients, here v15 and v27 (whose weights to ui ?s are shown at the axes extrema). For
example, the upper-left block contains image patches for which v 15 was highly negative
(contrast localized to bottom half of patch) and v27 was highly positive (power predominantly at low spatial scales). This illustrates how different combinations of basis functions
in B define distributions of images (in this case, spatial frequency and location).
Another way to get insight into the code learned by the model is to display, for a large
ensemble of image patches, the patches that yield the largest values of particular v i ?s (and
their corresponding basis functions in B). This is shown in fig. 4.
As a check to see if any of the higher-order structure learned by the algorithm was simply
due to random variations in the dataset, we generated a dataset by drawing independent
samples un from a generalized Gaussian to produce the pattern xn = Aun . The resulting
basis B was composed only of small random values, indicating essentially no deviation
from the standard assumption of independence and unit variance. In addition, adapting
the model on a synthetic dataset generated from a hand-specified B recovers the original
higher-order basis functions.
It is also possible to adapt A and B simultaneously (although with considerably greater
computational expense). To check the validity of first deriving B for a fixed A, both matrices were adapted simultaneously for small 8 ? 8 patches on the same natural image data
set. The results for both the image basis matrix A and the higher-order basis B were qualitatively similar to those reported above.
4 Discussion
We have presented a model for learning higher-order statistical regularities in natural images by learning an efficient, sparse-distributed code for the basis function coefficient variances. The recognition algorithm is non-linear, but we have not tested yet whether it can
account for non-linearities similar to the types reported in [10].
A (cautious) neurobiological interpretation of the higher-order units is that they are analogous to complex cells which pool output over specific first-order feature dimensions.
Rather than achieving a simplistic invariance, however, the model presented here has the
specific goal of efficiently representing the higher-order structure by adapting to the statistics of natural images, and thus may predict a broader range of response properties than are
commonly tested physiologically.
One salient type of higher-order structure learned by the model is the position of image
structure within the patch. It is interesting that, rather than encoding specific locations, the
model learned a coarse code of position using broadly tuned spatial patterns. This could
offer novel insights into the function of the broad tuning of higher level visual neurons.
By learning higher-order basis functions for different classes of visual images, the model
could not only provide insights into other types of visual response properties, but could
provide a way to simplify some of the computations in perceptual organization and other
computations in mid-level vision.
References
[1] H. B. Barlow. Possible principles underlying the transformation of sensory messages. In W. A.
Rosenbluth, editor, Sensory Communication, pages 217?234. MIT Press, Cambridge, 1961.
[2] B. A. Olshausen and D. J. Field. Emergence of simple-cell receptive-field properties by learning
a sparse code for natural images. Nature, 381:607?609, 1996.
[3] A. J. Bell and T. J. Sejnowski. The ?independent components? of natural scenes are edge filters.
Vision Res., 37(23):3327?3338, 1997.
[4] J. H. van Hateren and A. van der Schaaf. Independent component filters of natural images
compared with simple cells in primary visual cortex. Proc. Royal Soc. Lond. B, 265:359?366,
1998.
[5] J. H. van Hateren and D. L. Ruderman. Independent component analysis of natural image
sequences yield spatiotemporal filters similar to simple cells in primary visual cortex. Proc.
Royal Soc. Lond. B, 265:2315?2320, 1998.
[6] P. O. Hoyer and A. Hyvarinen. Independent component analysis applied to feature extraction
from colour and stereo images. Network, 11(3):191?210, 2000.
[7] E. Simoncelli and B. Olshausen. Natural image statistics and neural representation. Ann. Rev.
Neurosci., 24:1193?1216, 2001.
[8] T-W. Lee and M. S. Lewicki. Unsupervised classification, segmentation and de-noising of
images using ICA mixture models. IEEE Trans. Image Proc., 11(3):270?279, 2002.
[9] P. O. Hoyer and A. Hyvarinen. A multi-layer sparse coding network learns contour coding from
natural images. Vision Research, 42(12):1593?1605, 2002.
[10] O. Schwartz and E. P. Simoncelli. Natural signal statistics and sensory gain control. Nat.
Neurosci., 4:819?825, 2001.
[11] B. A. Pearlmutter and L. C. Parra. A context-sensitive generalization of ICA. In International
Conference on Neural Information Processing, pages 151?157, 1996.
[12] J-F. Cardoso. Infomax and maximum likelihood for blind source separation. IEEE Signal
Processing Letters, 4:109?111, 1997.
| 2167 |@word middle:1 simulation:1 contains:2 tuned:1 current:1 scatter:1 yet:1 realistic:1 shape:2 plot:1 half:2 inspection:1 beginning:1 ith:1 coarse:2 provides:3 location:7 fitting:1 cnbc:1 ica:8 themselves:2 multi:1 becomes:1 discover:1 linearity:3 underlying:1 maximizes:1 what:1 kind:2 extremum:1 transformation:3 schwartz:3 control:1 unit:2 omit:1 appear:2 positive:4 local:1 limit:2 encoding:2 approximately:2 might:2 black:2 au:1 equivalence:1 limited:6 range:8 statistically:1 v27:2 block:4 procedure:1 yan:2 bell:1 adapting:3 gabor:4 refers:1 get:1 cannot:1 noising:1 context:3 center:2 maximizing:3 go:1 straightforward:2 starting:1 simplicity:1 insight:4 deriving:1 embedding:1 notion:1 variation:3 analogous:1 us:1 hypothesis:1 element:1 approximated:1 ze:1 recognition:1 observed:2 bottom:4 capture:8 region:3 counter:1 ui:9 complexity:2 efficiency:3 basis:53 compactly:1 joint:7 represented:2 derivation:1 describe:5 sejnowski:1 apparent:1 whose:2 drawing:1 otherwise:1 statistic:6 emergence:1 sequence:1 propose:1 aun:1 product:3 adaptation:1 achieve:1 description:3 cautious:1 regularity:8 produce:1 object:7 derive:5 soc:2 c:1 drawback:1 filter:4 centered:2 generalization:2 probable:1 parra:1 around:1 exp:2 cognition:1 predict:1 major:2 early:1 proc:3 superposition:1 sensitive:1 largest:2 reflects:1 mit:1 clearly:1 gaussian:4 rather:2 broader:1 encode:1 derived:2 ax:1 likelihood:7 check:2 contrast:4 posteriori:1 factoring:1 pixel:3 among:9 orientation:6 classification:1 exponent:1 spatial:9 schaaf:1 marginal:2 field:2 construct:1 having:1 extraction:1 represents:2 broad:1 unsupervised:1 peaked:1 others:1 fundamentally:2 simplify:1 few:1 primarily:1 randomly:1 composed:1 simultaneously:2 argmax:3 organization:2 message:1 highly:3 predominant:1 mixture:2 light:1 edge:1 closer:1 necessary:1 orthogonal:1 initialized:1 re:1 minimal:1 fitted:1 column:4 modeling:3 v15:1 introducing:1 deviation:5 subset:2 uniform:1 graphic:1 reported:2 dependency:7 spatiotemporal:1 synthetic:1 considerably:1 density:2 fundamental:1 international:1 lee:1 physic:1 pool:1 michael:1 infomax:1 reflect:2 central:1 account:3 de:1 retinal:1 unordered:1 coding:12 coefficient:28 explicitly:1 vi:6 blind:1 performed:1 break:1 view:2 start:1 recover:1 square:1 variance:20 efficiently:3 ensemble:6 yield:4 correspond:2 bayesian:1 explain:2 against:1 frequency:2 obvious:1 associated:1 recovers:1 sampled:1 auditory:1 dataset:3 radially:1 gain:1 covariation:1 segmentation:2 nerve:1 higher:32 follow:1 response:5 maximally:1 formulation:2 arranged:2 correlation:3 hand:1 eqn:1 horizontal:1 ruderman:1 propagation:1 olshausen:2 validity:1 true:1 barlow:1 spatially:1 white:2 generalized:5 theoretic:1 complete:1 pearlmutter:1 image:69 novel:3 predominantly:1 common:2 extend:1 tail:1 interpretation:1 marginals:1 mellon:1 significant:1 refer:1 cambridge:1 automatic:1 tuning:1 dot:5 cortex:4 surface:4 inhibition:1 posterior:2 showed:1 optimizes:1 deta:1 der:1 greater:1 maximize:1 clockwise:1 signal:2 simoncelli:4 reduces:1 characterized:1 adapt:2 offer:4 coded:1 basic:2 simplistic:1 vision:5 cmu:2 essentially:1 iteration:1 represent:5 normalization:1 achieved:1 cell:7 invert:1 background:1 want:2 addition:1 addressed:1 source:1 envelope:1 ascent:4 tend:1 near:1 independence:9 affect:1 fit:1 det:2 whether:1 colour:1 nyquist:1 stereo:1 nine:2 useful:1 cardoso:1 mid:1 ten:1 simplest:1 rearranged:3 specifies:1 estimated:2 broadly:1 carnegie:1 redundancy:1 salient:2 achieving:1 clarity:1 inverse:1 letter:1 patch:16 separation:1 scaling:1 layer:2 display:1 correspondence:1 replaces:1 yielded:1 adapted:4 bv:6 scene:8 aspect:1 lond:2 formulating:1 performing:2 relatively:1 department:1 developing:1 according:4 combination:1 across:3 describes:2 rev:1 dv:1 computationally:1 visualization:1 remains:1 describing:1 needed:1 hierarchical:1 occurrence:1 alternative:1 batch:1 original:3 assumes:2 top:2 reflectance:1 arrangement:1 receptive:1 primary:4 unclear:1 exhibit:1 gradient:5 hoyer:2 higherorder:2 lateral:1 whom:1 toward:1 reason:1 assuming:1 code:12 modeled:2 relationship:3 insufficient:1 x20:1 unfortunately:1 expense:1 negative:4 rosenbluth:1 upper:1 observation:2 neuron:3 anti:1 sparsified:1 communication:1 dc:1 overcoming:1 inverting:1 specified:1 optimized:1 learned:9 trans:1 beyond:2 usually:1 pattern:12 below:2 challenge:2 royal:2 power:1 suitable:1 natural:25 karklin:1 representing:2 axis:1 deviate:3 prior:4 text:1 marginalizing:1 relative:1 loss:1 interesting:1 limitation:3 localized:2 degree:3 principle:2 editor:1 row:4 wide:3 emerge:1 sparse:9 distributed:3 van:3 default:3 xn:3 dimension:1 rich:1 contour:1 sensory:4 commonly:2 adaptive:1 qualitatively:1 hyvarinen:2 approximate:2 compact:1 implicitly:1 absorb:1 neurobiological:1 reveals:1 assumed:4 un:2 physiologically:1 learn:5 nature:2 complex:7 neurosci:2 x1:1 fig:11 precision:1 position:6 perceptual:1 learns:3 wavelet:1 down:1 specific:6 physiological:1 intractable:1 texture:4 nat:1 illustrates:1 simply:5 visual:16 contained:1 lewicki:3 determines:2 goal:4 formulated:1 ann:1 feasible:1 change:1 determined:2 called:1 invariance:1 divisive:1 meaningful:1 indicating:1 internal:1 hateren:2 tested:2 correlated:2 |
1,282 | 2,168 | Rate Distortion Function in the Spin Glass State:
a Toy Model
Tatsuto Murayama and Masato Okada
Laboratory for Mathematical Neuroscience
RIKEN Brain Science Institute
Saitama, 351-0198, JAPAN
{murayama,okada}@brain.riken.go.jp
Abstract
We applied statistical mechanics to an inverse problem of linear mapping
to investigate the physics of optimal lossy compressions. We used the
replica symmetry breaking technique with a toy model to demonstrate
Shannon?s result. The rate distortion function, which is widely known
as the theoretical limit of the compression with a fidelity criterion, is
derived. Numerical study shows that sparse constructions of the model
provide suboptimal compressions.
1
Introduction
Many information-science studies are very similar to those of statistical physics. Statistical
physics and information science may have been expected to be directed towards common
objectives since Shannon formulated an information theory based on the concept of entropy. However, envisaging how this actually happened would have been difficult; that the
physics of disordered systems, and spin glass theory in particular, at its maturity naturally
includes some important aspects of information sciences, thus reuniting the two disciplines.
This cross-disciplinary field can thus be expected to develop much further beyond current
perspectives in the future [1].
The areas where these relations are particularly strong are Shannon?s coding theory [2] and
classical spin systems with quenched disorder, which is the replica theory of disordered
statistical systems [3]. Triggered by the work of Sourlas [4], these links have recently
been examined in the area of matrix-based error corrections [5, 6], network-based compressions [7], and turbo decoding [8]. Recent results of these topics are mostly based on
the replica technique. Without exception, their basic characteristics (such as channel capacity, entropy rate, or achievable rate region) are only captured by the concept of a phase
transition with a first-order jump between the optimal and the other solutions arising in the
scheme.
However, the research in the cross-disciplinary field so far can be categorized as a so-called
?zero-distortion? decoding scheme in terms of information theory: the system requires perfect reproduction of the input alphabets [2]. Here, the same spin glass techniques should
be useful to describe the physics of systems with a fidelity criterion; i.e., a certain degree
of information distortion is assumed when reproducing the alphabets. This framework is
called the rate distortion theory [9, 10]. Though processing information requires regarding the concept of distortions practically, where input alphabets are mostly represented by
continuous variables, statistical physics only employs a few approaches [11, 12].
In this paper, we introduce a prototype that is suitable for cross-disciplinary study. We
analyze how information distortion can be described by the concepts of statistical physics.
More specifically, we study the inverse problem of a Sourlas-type decoding problem by using the framework of replica symmetry breaking (RSB) of diluted disordered systems [13].
According to our analysis, this simple model provides an optimal compression scheme
for an arbitrary fidelity-criterion degree, though the encoding procedure remains an NPcomplete problem without any practical encoders.
The paper is organized as follows. In Section 2, we briefly review the concept of the
rate distortion theory as well as the main results related to our purpose. In Section 3, we
introduce a toy model. In Section 4, we obtain consistent results with information theory.
Conclusions are given in the last section. Detailed derivations will be reported elsewhere.
2
Review: Rate Distortion Theory
We briefly recall the definitions of the concepts of the rate distortion theory and state the
simplest version of the main result at the end of this section. Let J be a discrete random variable with alphabet J . Assume that we have a source that produces a sequence
J1 , J2 , ? ? ? , JM , where each symbol is randomly drawn from a distribution. We will assume that the alphabet is finit. Throughout this paper we use vector notation to represent
sequences for convenience of explanation: J = (J1 , J2 , ? ? ? , JM )T ? J M . Here, the
encoder describes the source sequence J ? J M by a codeword ? = f (J ) ? X N . The
? = g(?) ? J?M , as illustrated in Figure 1. Note that
decoder represents J by an estimate J
M represents the length of a source sequence, while N represents the length of a codeword.
Here, the rate is defined by R = N/M . Note that the relation N < M always holds when
a compression is considered; therefore, R < 1 also holds.
Definition 2.1 A distortion function is a mapping
d : J ? J? ? R+
(1)
from the set of source alphabet-reproduction alphabet pairs into the set of non-negative
real numbers.
? is a measure of the cost of representing the symbol J by
Intuitively, the distortion d(J, J)
?
the symbol J. This definition is quite general. In most cases, however, the reproduction
alphabet J? is the same as the source alphabet J . Hereafter, we set J? = J and the
following distortion measure is adopted as the fidelity criterion:
Definition 2.2 The Hamming distortion is given by
(
?
? = 0 if J = J ,
d(J, J)
1 if J 6= J?
(2)
,
? = P[J 6= J]
?
which results in a probable error distortion, since the relation E[d(J, J)]
holds, where E[?] represents the expectation and P[?] the probability of its argument. The
distortion measure is so far defined on a symbol-by-symbol basis. We extend the definition
to sequences:
Definition 2.3 The distortion between sequences J , J? ? J M is defined by
M
X
?) = 1
d(Jj , J?j ) .
d(J , J
M j=1
(3)
Therefore, the distortion for a sequence is the average distortion per symbol of the elements
of the sequence.
Definition 2.4 The distortion associated with the code is defined as
? )] ,
D = E[d(J , J
(4)
where the expectation is with respect to the probability distribution on J .
A rate distortion pair (R, D) should be achiebable if a sequence of rate distortion codes
? )] ? D in the limit M ? ?. Moreover, the closure of the set
(f, g) exist with E[d(J , J
of achievable rate distortion pairs is called the rate distortion region for a source. Finally,
we can define a function to describe the boundary:
Definition 2.5 The rate distortion function R(D) is the infimum of rates R, so that (R, D)
is in the rate distortion region of the source for a given distortion D.
As in [7], we restrict ourselves to a binary source J with a Hamming distortion measure
for simplicity. We assume that binary alphabets are drawn randomly, i.e., the source is
not biased to rule out the possiblity of compression due to redundancy. We now find the
description rate R(D) required to describe the source with an expected proportion of errors
less than or equal to D. In this simplified case, according to Shannon, the boundary can be
written as follows.
Theorem 2.1 The rate distortion function for a binary source with Hamming distortion is
given by
1 ? H(D) 0 ? D ? 21
R(D) =
,
(5)
1
0
2 <D
where H(?) represents the binary entropy function.
encoder
decoder
?
J ?? f ?? ? ?? g ?? J
Figure 1: Rate distortion encoder and decoder
3
General Scenario
In this section, we introduce a toy model for lossy compression. We use the inverse problem
of Sourlas-type decoding to realize the optimal encoding scheme [4]. As in the previous
section, we assume that binary alphabets are drawn randomly from a non-biased source
and that the Hamming distortion measure is selected for the fidelity criterion.
We take the Boolean representation of the binary alphabet J , i.e., we set J = {0, 1}.
We also set X = {0, 1} to represent the codewords throughout the rest of this paper.
? an M -bit reproduction
Let J be an M -bit source sequence, ? an N -bit codeword, and J
sequence. Here, the encoding problem can be written as follows. Given a distortion D and
a randomly-constructed Boolean matrix A of dimensionality M ? N , we find the N -bit
codeword sequence ?, which satisfies
? = A?
J
(mod 2) ,
(6)
where the fidelity criterion
? )]
D = E[d(J , J
(7)
holds, according to every M -bit source sequence J . Note that we applied modulo 2 arithmetics for the additive operations in (6). In our framework, decoding will just be a linear
? = A?, while encoding remains a NP-complete problem.
mapping J
Kabashima and Saad recently expanded on the work of Sourlas, which focused on the zerorate limit, to an arbitrary-rate case [5]. We follow their construction of the matrix A, so we
can treat practical cases. Let the Boolean matrix A be characterized by K ones per row
and C per column. The finite, and usually small, numbers K and C define a particular
code. The rate of our codes can be set to an arbitrary value by selecting the combination
of K and C. We also use K and C as control parameters to define the rate R = K/C. If
the value of K is small, i.e., the relation K N holds, the Boolean matrix A results in
a very sparse matrix. By contrast, when we consider densely constructed cases, K must
be extensively big and have a value of O(N ). We can also assume that K is not O(1) but
K N holds. The codes within any parameter region, including the sparsely-constructed
cases, will result in optimal codes as we will conclude in the following section. This is one
new finding of our analysis using statistical physics.
4
Physics of the model: One-step RSB Scheme
The similarity between codes of this type and Ising spin systems was first pointed out by
Sourlas, who formulated the mapping of a code onto an Ising spin system Hamiltonian
in the context of error correction [4]. To facilitate the current investigation, we first map
the problem to that of an Ising model with finite connectivity following Sourlasfmethod.
We use the Ising representation {1, ?1} of the alphabet J and X rather than the Boolean
one {0, 1}; the elements of the source J and the codeword sequences ? are rewritten in
Ising values by mapping only, and the reproduction sequence J? is generated by taking
products of the relevant binary codeword sequence elements in the Ising representation
J?hi1 ,i2 ,??? ,iK i = ?i1 ?i2 ? ? ? ?iK , where the indices i1 , i2 , ? ? ? , iK correspond to the ones per
row A, producing a Ising version of J?. Note that the additive operation in the Boolean
representation is translated into the multiplication in the Ising one. Hereafter, we set
Jj , J?j , ?i = ?1 while we do not change the notations for simplicity. As we use statisticalmechanics techniques, we consider the source and codeword-sequence dimensionality (M
and N , respectively) to be infinite, keeping the rate R = N/M finite. To explore the
system?s capabilities, we examine the Hamiltonian:
X
H(S) = ?
Ahi1 ,??? ,iK i Jhi1 ,??? ,iK i Si1 ? ? ? SiK ,
(8)
hi1 ,??? ,iK i
where we have introduced the dynamical variable Si to find the most suitable Ising codeword sequence ? to provide the reproduction sequence J? in the decoding stage. Elements
of the sparse connectivity tensor Ahi1 ,??? ,iK i take the value one if the corresponding indices
of codeword bits are chosen (i.e., if all corresponding indices of the matrix A are one) and
zero otherwise; C ones per i index represent the system?s degree of connectivity.
For calculating the partition function Z(A, J ) = Tr{S} exp[??H(S)], we apply the
replica method following the calculation of Kabashima and Saad [5]. To calculate replicafree energy, we have to calculate the annealed average of the n-th power of the partition
function by preparing n replicas. Here we introduce the inverse temperature ?, which can
be interpreted as a measure of the system?s sensitivity to distortions. As we see in the following calculation, the optimal value of ? is naturally determined when the consistency of
the replica symmetry breaking scheme is considered [13, 3]. We use integral representations of the Dirac ? function to enforce the restriction, C bonds per index, on A [14]:
?
?
I 2?
X
dZ ?(C+1) Phi ,i ,??? ,i i Ahi,i2 ,??? ,iK i
K
??
Ahi,i2 ,??? ,iK i ? C ? =
Z
Z 2 3
, (9)
2?
0
hi2 ,i3 ,??? ,iK i
giving rise to a set of order parameters
q?,?,??? ,?
N
1 X
Zi Si? Si? ? ? ? Si? ,
=
N i=1
(10)
where ?, ?, ? ? ? , ? represent replica indices, and the average over J is taken with respect
to the probability distribution:
P[Jhi1 ,i2 ,??? ,iK i ] =
1
1
?(Jhi1 ,i2 ,??? ,iK i ? 1) + ?(Jhi1 ,i2 ,??? ,iK i + 1)
2
2
(11)
as we consider the non-biased source sequences for simplicity. Assuming the replica symmetry, we use a different representation for the order parameters and the related conjugate
variables [14]:
Z
q?,?,??? ,? = q dx ?(x) tanhl (?x) ,
(12)
Z
q??,?,??? ,? = q? dx ?
? (?
x) tanhl (? x
?) ,
(13)
where q = [(K ? 1)!N C]1/K and q? = [(K ? 1)!]?1/K [N C](K?1)/K are normalization
constants, and ?(x) and ?
? (?
x) represent probability distributions related to the integration
variables. Here l denotes the number of related replica indices. Throughout this paper,
integrals with unspecified limits denote integrals over the range of (??, +?). We then
obtain an expression for the free energy per source bit expressed in terms of the probability
distributions ?(x) and ?
? (?
x):
1
hhln Z(A, J )ii
M
= ln cosh ?
*
!+
Z Y
K
K
Y
+
dxl ?(xl ) ln 1 + tanh ?J
tanh ?xl
??f =
l=1
?K
Z
l=1
dx ?(x)
J
(14)
Z
d?
x?
? (?
x) ln(1 + tanh ?x tanh ? x
?)
"
#
Z Y
C
C
Y
C
?l ) ,
d?
xl ?
? (?
xl ) ln Tr (1 + S tanh ? x
+
S
K
l=1
l=1
where hh? ? ? ii denotes the average over quenched randomness of A and J . The saddle
point equations with respect to probability distributions provide a set of relations between
?(x) and ?
? (?
x):
#
!
Z "C?1
C?1
Y
X
?(x) =
d?
xl ?
? (?
xl ) ? x ?
x
?l ,
l=1
?
? (?
x) =
Z "C?1
Y
l=1
l=1
dxl ?(xl )
#* "
1
? x
? ? tanh?1
?
tanh ?J
K?1
Y
tanh ?xl
l=1
!#+
(15)
.
J
By using the result obtained for the free energy, we can easily perform further straightforward calculations to find all the other observable thermodynamical quantities, including
internal energy:
EE
1 DD
1 ?
TrS H(S)e??H(S)
hhln Z(A, J )ii ,
(16)
e=
=?
M
M ??
which records reproduction errors. Therefore, in terms of the considered replica symmetric
ansatz, a complete solution of the problem seems to be easily obtainable; unfortunately, it
is not.
This set of equations (15) may be solved numerically for general ?, K, and C. However, there exists an analytical solution of this equations. We first consider this case. Two
dominant solutions emerge that correspond to the paramagnetic and the spin glass phases.
The paramagnetic solution, which is also valid for general ?, K, and C, is in the form of
?(x) = ?(x) and ?
? = ?(?
x); it has the lowest possible free energy per bit fPARA = ?1,
although its entropy sPARA = (R?1) ln 2 is positive only for R ? 1. It means that the true
solution must be somewhere beyond the replica symmetric ansatz. As a first step, which is
called the one-step replica symmetry breaking (RSB), n replicas are usually divided into
n/m groups, each containing m replicas. Pathological aspects due to the replica symmetry
may be avoided making use of the newly-defined freedom m. Actually, this one-step RSB
scheme is considered to provide the exact solutions when the random energy model limit
is considered [15], while our analysis is not restricted to this case.
The spin glass solution can be calculated for both the replica symmetric and the one-step
RSB ansatz. The former reduces to the paramagnetic solution (fRS = fPARA ), which is
unphysical for R < 1, while the latter yields ?1RSB (x) = ?(x), ?
?1RSB (?
x) = ?(?
x) with
m = ?g (R)/? and ?g obtained from the root of the equation enforcing the non-negative
replica symmetric entropy
sRS = ln cosh ?g ? ?g tanh ?g + R ln 2 = 0 ,
(17)
with a free energy
f1RSB = ?
1
R
ln cosh ?g ?
ln 2 .
?g
?g
(18)
Since the target bit of the estimation in this model is Jhi1 ,??? ,iK i and its estimator the product
Si1 ? ? ? SiK , a performance measure for the information corruption could be the per-bond
energy e. According to the one-step RSB framework, the lowest free energy can be calculated from the probability distributions ?1RSB (x) and ?
?1RSB (?
x) satisfying the saddle point
equation (15) at the characteristic inverse temperature ?g , when the replica symmetric entropy sRS disappears. Therefore, f1RSB equals e1RSB . Let the Hamming distortion be our
fidelity criterion. The distortion D associated with this code is given by the fraction of the
free energies that arise in the spin glass phase:
D=
f1RSB ? fRS
1 ? tanh ?g
=
.
2|fRS |
2
(19)
Here, we substitute the spin glass solutions into the expression, making use of the fact that
the replica symmetric entropy sRS disappears at a consistent ?g , which is determined by
(17). Using (17) and (19), simple algebra gives the relation between the rate R = N/M
and the distortion D in the form
R = 1 ? H(D) ,
which coincides with the rate distortion function retrieving Theorem 2.1. Surprisingly, we
do not observe any first-order jumps between analytical solutions. Recently, we have seen
that many approaches to the family of codes, characterized by the linear encoding operations, result in a quite different picture: the optimal boundary is constructed in the random
energy model limit and is well captured by the concept of a first-order jump. Our analysis
of this model, viewed as a kind of inverse problem, provides an exception. Many optimal
conditions in textbook information theory may be well described without the concept of a
first-order phase transitions from a view point of statistical physics.
We will now investigate the possiblity of the other solutions satisfying (15) in the case
of finite K and C. Since the saddle point equations (15) appear difficult for analytical
arguments, we resort to numerical evaluations representing the probability distributions
?1RSB (x) and ?
?1RSB (?
x) by up to 105 bin models and carrying out the integrations by
using Monte Carlo methods. Note that the characteristic inverse temperature ? g is also
evaluated numerically by using (17). We set K = 2 and selected various values of C to
demonstrate the performance of stable solutions. The numerical results obtained by the
one-step RSB senario show suboptimal properties [Figure 2]. This strongly implies that
the analytical solution is not the only stable solution. This conjecture might be verified
elsewhere, carrying out large scale simulations.
5
Conclusions
Two points should be noted. Firstly, we found that the consistency between the rate distortion theory and the Parisi one-step RSB scheme. Secondly, we conjectured that the
analytical solution, which is consistent with the Shannon?s result, is not the only stable
solution for some situations. We are currently working on the verification.
Acknowledgments
We thank Yoshiyuki Kabashima and Shun-ichi Amari for their comments on the
manuscript. We also thank Hiroshi Nagaoka and Te Sun Han for giving us valuable references. This research is supported by the Special Postdoctoral Researchers Program at
RIKEN.
References
[1] H. Nishimori. Statistical Physics of Spin Glasses and Information Processing. Oxford
University Press, 2001.
[2] T. M. Cover and J. A. Thomas. Elements of Information Theory. Wiley, 1991.
[3] V. Dotsenko. Introduction to the Replica Theory of Disordered Statistical Systems.
Cambridge University Press, 2001.
[4] N. Sourlas. Spin-glass models as error-correcting codes. Nature, 339:693?695, 1989.
[5] Y. Kabashima and D. Saad. Statistical mechanics of error-correcting codes. Europhys.
Lett., 45:97?103, 1999.
[6] Y. Kabashima, T. Murayama, and D. Saad. Typical performance of Gallager-type
error-correcting codes. Phys. Rev. Lett., 84:1355?1358, 2000.
1
K=2
R(D)
2
0.8
^ )
?^ ( x
1
R
0.6
? ( x )
0.4
0
-2
-1
0
1
2
0.2
0
0
0.1
0.2
0.3
0.4
0.5
D
Figure 2: Numerically-constructed stable solutions: Stable solutions of (15) for the finite
values of K and L are calculated by using Monte Carlo methods. We use 105 bin models
to approximate the probability distributions ?1RSB (x) and ?
?1RSB (?
x), starting from various
initial conditions. The distributions converge to the continuous ones, giving suboptimal
performance. (?) K = 2 and L = 3, 4, ? ? ? , 12 ; Solid line indicates the rate distortion
function R(D). Inset: Snapshots of the distributions, where L = 3 and ?g = 2.35.
[7] T. Murayama. Statistical mechanics of the data compression theorem. J. Phys. A,
35:L95?L100, 2002.
[8] A. Montanari and N. Sourlas. The statistical mechanics of turbo codes. Eur. Phys. J.
B, 18:107?119, 2000.
[9] C. E. Shannon. Coding theorems for a discrete source with a fidelity criterion. IRE
National Convention Record, Part 4, pages 142?163, 1959.
[10] T. Berger. Rate Distortion Theory: A Mathematical Basis for Data Compression.
Prentice-Hall, 1971.
[11] T. Hosaka, Y. Kabashima, and H. Nishimori. Statistical mechanics of lossy data
compression using a non-monotonic perceptron. cond-mat/0207356.
[12] Y. Matsunaga and H. Yamamoto. A coding theorem for lossy data compression by
LDPC codes. In Proceedings 2002 IEEE International Symposium on Information
Theory, page 461, 2002.
[13] M. Mezard, G. Parisi, and M. Virasoro. Spin-Glass Theory and Beyound. World
Scientific, 1987.
[14] K. Y. M. Wong and D. Sherrington. Graph bipartitioning and spin glasses on a random
network of fixed finite valence. J. Phys. A, 20:L793?L799, 1987.
[15] B. Derrida. The random energy model, an exactly solvable model of disordered systems. Phys. Rev. B, 24:2613?2626, 1981.
| 2168 |@word version:2 briefly:2 achievable:2 compression:12 proportion:1 seems:1 closure:1 simulation:1 tr:2 solid:1 initial:1 hereafter:2 selecting:1 current:2 paramagnetic:3 si:4 dx:3 written:2 must:2 realize:1 numerical:3 additive:2 j1:2 partition:2 selected:2 hamiltonian:2 record:2 provides:2 ire:1 firstly:1 si1:2 mathematical:2 constructed:5 symposium:1 maturity:1 ik:14 retrieving:1 introduce:4 expected:3 examine:1 mechanic:5 brain:2 jm:2 notation:2 moreover:1 lowest:2 kind:1 interpreted:1 unspecified:1 textbook:1 finding:1 every:1 exactly:1 control:1 appear:1 producing:1 positive:1 treat:1 limit:6 encoding:5 oxford:1 might:1 examined:1 range:1 directed:1 practical:2 acknowledgment:1 procedure:1 area:2 quenched:2 convenience:1 onto:1 prentice:1 context:1 senario:1 wong:1 restriction:1 map:1 dz:1 annealed:1 go:1 straightforward:1 starting:1 focused:1 bipartitioning:1 simplicity:3 disorder:1 correcting:3 rule:1 estimator:1 construction:2 target:1 modulo:1 exact:1 element:5 satisfying:2 particularly:1 sparsely:1 ising:9 solved:1 calculate:2 region:4 sun:1 valuable:1 carrying:2 algebra:1 basis:2 finit:1 translated:1 easily:2 represented:1 various:2 riken:3 alphabet:13 derivation:1 describe:3 monte:2 hiroshi:1 europhys:1 quite:2 widely:1 distortion:41 otherwise:1 amari:1 encoder:3 nagaoka:1 triggered:1 sequence:20 parisi:2 analytical:5 product:2 fr:3 j2:2 sourlas:7 relevant:1 murayama:4 description:1 hi2:1 dirac:1 produce:1 perfect:1 diluted:1 develop:1 derrida:1 strong:1 implies:1 convention:1 disordered:5 disciplinary:3 bin:2 shun:1 investigation:1 probable:1 secondly:1 correction:2 hold:6 practically:1 considered:5 hall:1 exp:1 mapping:5 purpose:1 dxl:2 estimation:1 bond:2 tanh:10 currently:1 always:1 i3:1 rather:1 derived:1 indicates:1 contrast:1 glass:11 relation:6 i1:2 fidelity:8 integration:2 special:1 field:2 equal:2 preparing:1 represents:5 future:1 np:1 employ:1 few:1 randomly:4 pathological:1 densely:1 national:1 phase:4 ourselves:1 freedom:1 investigate:2 evaluation:1 dotsenko:1 integral:3 yamamoto:1 theoretical:1 virasoro:1 column:1 boolean:6 cover:1 cost:1 saitama:1 reported:1 encoders:1 eur:1 international:1 sensitivity:1 physic:11 decoding:6 discipline:1 ansatz:3 connectivity:3 containing:1 resort:1 toy:4 japan:1 coding:3 includes:1 root:1 view:1 analyze:1 capability:1 spin:14 characteristic:3 who:1 correspond:2 yield:1 carlo:2 researcher:1 corruption:1 kabashima:6 randomness:1 phys:5 definition:8 energy:12 naturally:2 associated:2 hamming:5 newly:1 recall:1 dimensionality:2 organized:1 obtainable:1 actually:2 manuscript:1 follow:1 evaluated:1 though:2 strongly:1 just:1 stage:1 working:1 infimum:1 scientific:1 lossy:4 facilitate:1 concept:8 true:1 former:1 symmetric:6 laboratory:1 i2:8 illustrated:1 noted:1 coincides:1 criterion:8 complete:2 demonstrate:2 sherrington:1 npcomplete:1 temperature:3 recently:3 common:1 jp:1 extend:1 numerically:3 cambridge:1 consistency:2 pointed:1 stable:5 han:1 similarity:1 dominant:1 recent:1 perspective:1 conjectured:1 scenario:1 codeword:9 certain:1 binary:7 captured:2 seen:1 converge:1 arithmetic:1 ii:3 reduces:1 characterized:2 calculation:3 cross:3 divided:1 basic:1 expectation:2 represent:5 normalization:1 yoshiyuki:1 source:19 biased:3 rest:1 saad:4 sr:3 comment:1 mod:1 matsunaga:1 ee:1 zi:1 restrict:1 suboptimal:3 regarding:1 prototype:1 masato:1 possiblity:2 expression:2 jj:2 useful:1 detailed:1 extensively:1 cosh:3 simplest:1 exist:1 happened:1 neuroscience:1 arising:1 per:9 discrete:2 mat:1 group:1 redundancy:1 ichi:1 drawn:3 hi1:2 verified:1 l793:1 replica:21 graph:1 fraction:1 inverse:7 throughout:3 family:1 bit:9 turbo:2 trs:1 aspect:2 argument:2 expanded:1 conjecture:1 according:4 combination:1 conjugate:1 describes:1 rev:2 making:2 intuitively:1 restricted:1 taken:1 ln:9 equation:6 remains:2 hh:1 end:1 adopted:1 operation:3 rewritten:1 apply:1 observe:1 enforce:1 substitute:1 thomas:1 denotes:2 calculating:1 somewhere:1 giving:3 classical:1 tensor:1 objective:1 quantity:1 codewords:1 valence:1 link:1 thank:2 capacity:1 decoder:3 topic:1 enforcing:1 assuming:1 length:2 code:15 index:7 berger:1 ldpc:1 difficult:2 mostly:2 unfortunately:1 negative:2 rise:1 perform:1 snapshot:1 finite:6 situation:1 reproducing:1 arbitrary:3 introduced:1 pair:3 required:1 beyond:2 usually:2 dynamical:1 program:1 including:2 rsb:16 explanation:1 power:1 suitable:2 solvable:1 representing:2 scheme:8 picture:1 disappears:2 review:2 multiplication:1 nishimori:2 degree:3 verification:1 consistent:3 sik:2 dd:1 row:2 elsewhere:2 surprisingly:1 last:1 keeping:1 free:6 supported:1 perceptron:1 institute:1 taking:1 emerge:1 sparse:3 boundary:3 calculated:3 lett:2 transition:2 valid:1 world:1 jump:3 simplified:1 avoided:1 far:2 approximate:1 observable:1 assumed:1 conclude:1 postdoctoral:1 continuous:2 channel:1 okada:2 nature:1 symmetry:6 ahi:2 main:2 montanari:1 big:1 arise:1 categorized:1 wiley:1 mezard:1 xl:8 breaking:4 theorem:5 inset:1 symbol:6 reproduction:7 exists:1 te:1 entropy:7 explore:1 saddle:3 gallager:1 expressed:1 phi:1 monotonic:1 satisfies:1 viewed:1 formulated:2 towards:1 change:1 specifically:1 infinite:1 determined:2 typical:1 unphysical:1 called:4 shannon:6 cond:1 exception:2 internal:1 reuniting:1 latter:1 |
1,283 | 2,169 | Feature Selection by Maximum Marginal
Diversity
Nuno Vasconcelos
Department of Electrical and Computer Engineering
University of California, San Diego
[email protected]
Abstract
We address the question of feature selection in the context of visual
recognition. It is shown that, besides efficient from a computational
standpoint, the infomax principle is nearly optimal in the minimum
Bayes error sense. The concept of marginal diversity is introduced, leading to a generic principle for feature selection (the principle of maximum
marginal diversity) of extreme computational simplicity. The relationships between infomax and the maximization of marginal diversity are
identified, uncovering the existence of a family of classification procedures for which near optimal (in the Bayes error sense) feature selection
does not require combinatorial search. Examination of this family in light
of recent studies on the statistics of natural images suggests that visual
recognition problems are a subset of it.
1 Introduction
It has long been recognized that feature extraction and feature selection are important problems in statistical learning. Given a classification or regression task in some observation
space (typically high-dimensional), the goal is to find the best transform into a feature
space (typically lower dimensional) where learning is easier (e.g. can be performed with
less training data). While in the case of feature extraction there are few constraints on ,
for feature selection the transformation is constrained to be a projection, i.e. the components of a feature vector in are a subset of the components of the associated vector in .
Both feature extraction and selection can be formulated as optimization problems where
the goal is to find the transform that best satisfies a given criteria for ?feature goodness?.
In this paper we concentrate on visual recognition, a subset of the classification problem for
which various optimality criteria have been proposed throughout the years. In this context,
the best feature spaces are those that maximize discrimination, i.e. the separation between
the different image classes to recognize. However, classical discriminant criteria such as
linear discriminant analysis make very specific assumptions regarding class densities, e.g.
Gaussianity, that are unrealistic for most problems involving real data. Recently, various
authors have advocated the use of information theoretic measures for feature extraction or
selection [15, 3, 9, 11, 1]. These can be seen as instantiations of the the infomax principle
of neural organization1 proposed by Linsker [7], which also encompasses information theoretic approaches for independent component analysis and blind-source separation [2]. In
the classification context, infomax recommends the selection of the feature transform that
maximizes the mutual information (MI) between features and class labels.
While searching for the features that preserve the maximum amount of information about
the class is, at an intuitive level, an appealing discriminant criteria, the infomax principle
does not establish a direct connection to the ultimate measure of classification performance
- the probability of error (PE). By noting that to maximize MI between features and class
labels is the same as minimizing the entropy of labels given features, it is possible to establish a connection through Fano?s inequality: that class-posterior entropy (CPE) is a lower
bound on the PE [11, 4]. This connection is, however, weak in the sense that there is little
insight on how tight the bound is, or if minimizing it has any relationship to minimizing
PE. In fact, among all lower bounds on PE, it is not clear that CPE is the most relevant.
An obvious alternative is the Bayes error (BE) which 1) is the tightest possible classifierindependent lower-bound, 2) is an intrinsic measure of the complexity of the discrimination
problem and, 3) like CPE, depends on the feature transformation and class labels alone.
Minimizing BE has been recently proposed for feature extraction in speech problems [10].
The main contribution of this paper is to show that the two strategies (infomax and minimum BE) are very closely related. In particular, it is shown that 1) CPE is a lower bound
on BE and 2) this bound is tight, in the sense that the former is a good approximation to the
latter. It follows that infomax solutions are near-optimal in the minimum BE sense. While
for feature extraction both infomax and BE appear to be difficult to optimize directly, we
show that infomax has clear computational advantages for feature selection, particularly in
the context of the sequential procedures that are prevalent in the feature selection literature [6]. The analysis of some simple classification problems reveals that a quantity which
plays an important role in infomax solutions is the marginal diversity: the average distance
between each of the marginal class-conditional densities and their mean. This serves as
inspiration to a generic principle for feature selection, the principle of maximum marginal
diversity (MMD), that only requires marginal density estimates and can therefore be implemented with extreme computational simplicity. While heuristics that are close to the MMD
principle have been proposed in the past, very little is known regarding their optimality.
In this paper we summarize the main results of a theoretical characterization of the problems for which the principle is guaranteed to be optimal in the infomax sense (see [13] for
further details). This characterization is interesting in two ways. First, it shows that there is
a family of classification problems for which a near-optimal solution, in the BE sense, can
be achieved with a computational procedure that does not involve combinatorial search.
This is a major improvement, from a computational standpoint, to previous solutions for
which some guarantee of optimality (branch and bound search) or near optimality (forward
or backward search) is available [6]. Second, when combined with recent studies on the
statistics of biologically plausible image transformations [8, 5], it suggests that in the context of visual recognition, MMD feature selection will lead to solutions that are optimal in
the infomax sense. Given the computational simplicity of the MMD principle, this is quite
significant. We present experimental evidence in support of these two properties of MMD.
2 Infomax vs minimum Bayes error
In this section we show that, for classification problems, the infomax principle is closely
related to the minimization of Bayes error. We start by defining these quantities.
1
Under the infomax principle, the optimal organization for a complex multi-layered perceptual
system is one where the information that reaches each layer is processed so that the maximum amount
of information is preserved for subsequent layers.
Theorem 1 Given a classification problem with
classes in a feature space , the decision function which minimizes the probability of classification error is the Bayes classifier
!#" , where $ is a random variable that assigns to one of
classes, and &%('*)+-,.,-,#+
/ . Furthermore, the PE is lower bounded by the Bayes error
0
1)32543687 9 !#" ;:;+
(1)
where 436 means expectation with respect to .
Proof: All proofs are omitted due to space considerations. They can be obtained by contacting the author.
Principle 1 (infomax) Consider an
-class classification problem with observations
@?
drawn from random variable < %
, and the set of feature transformations >=
. The best feature space is the one that maximizes the mutual information A $CBD
< , and A $EBFD G
where $ is the class indicator
variable defined above, D
H IKJ ML !N+FOPRQ*1S.TV6U W9Y X 6 L ZYZY#[
the mutual information between D and $ .
S.T9X S.W9X
]^ $ _2`]^ $ " D , where ]^ D \
It is straightforward to show that A D + $ \
2 IKJ ! PRQ* J [ is the entropy of D . Since the class entropy ]a $ does not depend on , infomax is equivalent to the minimization of the CPE ]^ $ " D . We next derive
a bound that plays a central role on the relationship between this quantity and BE.
Lemma 1H Consider a probability mass function b
)+jif and #J
1) . Then,
c' J d +-,.,.,.+ Jfe / such that g5h J h
)
PZQVnm
2o)p
j)325 J ;lk
]^ b N2
)
(2)
ZP Q
PRQ*
q
H
.J PZQ J . Furthermore, the bound is tight in the sense that equality
where ]a b r
s2
holds when
)
+Oixwz
u
y {|,
J t
m
(3)
2u) and J v
m
2o)
The following theorem follows from this bound.
Theorem 2 The BE of an -class classification problem with feature space
indicator variable $ , is lower bounded by
0 }
lk
PRQ*
)
]^ $ " D 2
PZQfnm
PZQ
2o)p
q
)+
%
where
is the random vector
from which
d features are drawn. When
?D ~
0
) this bound reduces to } lk ? ?F? e ]^ $ " D .
(
and class
(4)
is large
It is interesting
to note the relationship between (4) and Fano?s lower bound on the PE
?uk ? ?? d e ]^ $ " D r2 ? ?? d e . The two bounds are equal up to an additive constant
d
?e
( ? ?? e PRQ*?
? eE?xd ) that quickly decreases to zero with the number of classes . It follows
that, at least when the number of classes is large, Fano?s is really a lower bound on BE, not
only on PE. Besides making this clear, Theorem 2 is a relevant contribution in two ways.
First, since constants do not change the location of the bound?s extrema, it shows that infomax minimizes a lower bound on BE. Second, unlike Fano?s bound, it sheds considerable
insight on the relationship between the extrema of the bound and those of the BE.
In fact, it is clear from the derivation of the theorem that, the only reason why the righthand (RHS) and left-hand (LHS) sides of (4) differ is the application of (2). Figure 1
0.6
1
1
0.9
0.9
0.8
0.8
0.7
0.7
0.6
0.6
1?max(p)
H(p) ? log(3) + 1
0.2
p2
0
p2
0.4
0.5
0.5
?0.2
?0.4
0.4
0.4
0.3
0.3
0.2
0.2
?0.6
0.1
?0.8
0
0.1
0.2
0.3
0.4
0.5
p
0.6
0.7
0.8
0.9
0
1
0.1
0
0.1
0.2
0.3
1
0.4
0.5
p1
0.6
0.7
0.8
0.9
0
1
0
0.1
0.2
0.3
0.4
0.5
p1
0.6
0.7
0.8
0.9
1
Figure 1: Visualization of (2). Left: LHS and RHS versus for
. Middle:
contours of the LHS versus
! !
. Right: same, for RHS.
for
3
0.5
1
0.4
0.8
0.3
0.6
2
0.1
?y
L
*
H(Y|X)
1
0.2
?1
0.2
0
5
0
0.4
?2
0
5
5
5
?3
0
0
0
?y
?5
?5
?x
0
?y
?5
?5
?3
?x
?2
?1
0
?
1
2
3
x
Figure
"$#% & 2: The LHS of (4)"$as
# % & an approximation to (1) for a two-class Gaussian problem where
('*) ,+.-/
10324 and
('5) 6 +-7
89:2; . All plots are functions of 8 . Left: surface plot
of (1). Middle: surface plot of the LHS of (4). Right: contour plots of the two functions.
%`'m+< / , illustrating three
shows plots of the RHS and LHS of this equation when
interesting properties. First, bound (2) is tight in the sense defined in the lemma. Second,
the maximum of the LHS is co-located with that of the RHS. Finally, (like the RHS) the
LHS is a concave function of b and increasing (decreasing) when the RHS is. Due to these
properties, the LHS is a good approximation to the RHS and, consequently, the LHS of (4)
a good approximation to its RHS. It follows that infomax solutions will, in general, be very
similar to those that minimize the BE . This is illustrated by a simple example in Figure 2.
3 Feature selection
For feature extraction, both infomax and minimum BE are complicated problems that can
only be solved up to approximations [9, 11, 10]. It is therefore not clear which of the
two strategies will be more useful in practice. We now show that the opposite holds for
feature selection, where the minimization of CPE is significantly simpler than that of BE.
We start by recalling that, because the possible number of feature subsets in a feature
selection problem is combinatorial, feature selection techniques rely on sequential search
methods [6]. These methods proceed in a sequence of steps, each adding a set of features
to the current best subset, with the goal of optimizing a given cost function 2. We denote the
current subset by D7= , the added features by D7> and the new subset by D/?
D/> + D/= .
Theorem 3 Consider an -class classification problem with? observations drawn from a
random variable < %
, and a feature transformation =
. is a infomax feature
2
These methods are called forward search techniques. There is also an alternative set of backward
search techniques, where features are successively removed from an initial set containing all features.
We ignore the latter for simplicity, even though all that is said can be applied to them as well.
space if and only if i
0
y
K!&" j."Z"
k 0
K!&" O-"R"
(5)
H
< ,D
< , j
!j !j
where D
6 Y expectation with respect
0 7 J "Z" p:
I J ! PZQ denotes
S X 6 Y [ is the Kullback-Leibler
to the prior class probabilities and
X
divergence between J and . Furthermore, if D7?
D7> + D/= , the infomax cost function
decouples into two terms according to
0 9
? " O-"R"
! ?
0 9 L " +Fj."Z" 9
> =
0 9. ! " j."Z" ,
=
=
q
! > " =
(6)
Equation (5) exposes the discriminant nature of the infomax criteria. Noting that
!
K!&" j , it clearly favors feature spaces where each class-conditional
density is as distant as possible (in the KL sense) from the average among all classes.
This is a sensible way to quantify the intuition that optimal discriminant transforms are the
ones that best separate the different classes. Equation (6), in turn, leads to an optimal rule
with the current optimal solution D = : the set which
for finding the
0 features
9 D L > to > merge
" = +Fj."Z" 9 ! > " = . The equation also leads to a
minimizes
straightforward procedure for updating the optimal cost once this set is determined. On the
other hand, when the cost function is BE, the equivalent expression is
L p! > " + =
l !#" = O:"! , (7)
! > " =
Note that the non-linearity introduced by the operator, makes it impossible to ex !#" = ;: . For this reason,
press 4 7 l
8!#" ? ;: as a function of 4 7 *9
4
7
!#" ? O:8
4 4r
7
infomax is a better principle for feature selection problems than direct minimization of BE.
4 Maximum marginal diversity
To gain some intuition for infomax solutions, we next consider the Gaussian problem of
Figure 3. Assuming that the two classes have equal prior probabilities j)
m
)$#m , the marginals &%('- r*)9" )p and +%
'# r*)" m* are equal and feature , d does not contain any useful information for classification. On the other hand, because the classes are
clearly separated along the ) ? axis, feature , ? contains all the information available for
discriminating between them. The different discriminating powers of the two variables
are reflected
the infomax costs: while % ' *)V
+%
'# r*)9" )p
-%
'. r.)9" m leads
0 7 +%
by
'
K
*)" j."Z" % ' *)fO:/
%&0 .)f5
y -% 0 *)" ) (
c
y +% 0 r*)" m* it
to
0 7 +% 0 *)" j-"R" %&0 *)Vg ;,:from
g , and (5) recommends the selection of
follows that
, ? . This is unlike energy-based criteria, such as principal component analysis, that would
select , d . The key advantage of infomax is that it emphasizes marginal diversity.
Definition 1 Consider a classification problem on a feature space , and a random vector
feature vectors are drawn. Then, 132 , v o
0 D 7 +%(
4 ., )9" d j.+."Z,-" ,.,#% + ,4 *?)V ;:+from
iswhich
the marginal diversity of feature , v .
The intuition conveyed by the example above can be easily transformed into a generic
principle for feature selection.
Principle 2 (Maximum marginal diversity) The best solution for a feature selection
problem is to select the subset of features that leads to a set of maximally diverse marginal
densities.
2.5
2
0.025
PX |Y(x|1)
1
(x|2)
P
PX
|Y
PX
|Y
(x|1)
2
(x|2)
2
X |Y
1
1.5
2
0.02
1
0.5
1.5
x2
0.015
0
?0.5
0.01
1
0.005
0.5
?1
?1.5
?2
?5
?4
?3
?2
?1
0
x1
1
2
3
4
0
?50
5
?40
?30
?20
Figure 3: Gaussian problem with two classes
Left: contours of
?10
0
x
10
20
30
40
0
?2
50
?1.5
?1
?0.5
0
x
0.5
1
1.5
2
6 , in. Right:
the two-dimensions,
marginals for
.
probability. Middle: marginals for
.
This principle has two attractive properties. First it is inherently discriminant, recommending the elimination of the dimensions along which the projections of the class densities are
most similar. Second, it is straightforward to implement with the following algorithm.
Algorithm 1 (MMD feature selection) For a classification problem with features D
, d +.,.,-,.+ ,? ,
classes $ % ' )+.,-,.,#+
/ and class priors jE
J the following
procedure returns the top MMD features.
- foreach feature w % ' )+-,.,.,.+ / :
'*)+-,.,-,#+
v L of -% 4 *)" j ,
* foreach class &%(
d H v L
/ , compute an histogram estimate
v
e
* compute
,
H .J v L PRQ*| v L , # v
* compute the marginal diversity 1 2 , v
, where both the
and division , # are performed element-wise,
- order the features by decreasing diversity, i.e. find 'w d +-,.,-,.+w ? / such that
1 2 , v Mk 1 2 , v ' , and return ' , v ' +.,-,.,-+ , v / .
In general, there are no guarantees that MMD will lead to the infomax solution. In [13]
we seek a precise characterization of the problems where MMD is indeed equivalent to
infomax. Due to space limitations we present here only the main result of this analysis,
see [13] for a detailed derivation.
Theorem 4 Consider a classification problem with class labels drawn from a random
, d +.,.,-,.+ , ? and let
variable $ and features drawn from a random vector D
r
+
.
,
,
#
,
+
D
, d
, be the optimal feature subset of size in the infomax sense. If
? d
A , v BD d L v ?8d " $ +ji8wE%(' )+-,.,., + /
A , v B D d L v 8
? d / , the set D is also the optimal subset of size
where D d L v ?8d
\' , d +.,.,-,#+ , v x
MMD sense. Furthermore,
0
9
r!&" O-"R"
!
! 1
v" d
2 ,
v ,
(8)
in the
(9)
The theorem states that the MMD and infomax solutions will be identical when the mutual
information between features is not affected by knowledge of the class label. This is an
interesting condition in light of various recent studies that have reported the observationof
consistent patterns of dependence between the features of various biologically plausible
image transformations [8, 5]. Even though the details of feature dependence will vary from
one image class to the next, these studies suggest that the coarse structure of the patterns of
dependence between such features follow universal statistical laws that hold for all types of
images. The potential implications of this conjecture are quite significant. First it implies
1
1
2.2
0.95
1.8
2
0.95
0.8
Cumulative marginal diversity
Classification rate
Jain/Zongker score
0.9
0.85
0.9
0.85
1.6
1.4
1.2
1
0.8
0.75
DCT
PCA
Wavelet
0.8
0.7
DCT
PCA
Wavelet
0.6
0.4
0.65
1
10
2
3
10
10
4
10
Sample size
a)
0.75
0
5
10
15
20
Number of features
25
0.2
30
0
5
10
b)
15
20
Number of features
25
30
35
c)
Figure 4: a) JZ score as a function of sample size for the two-class Gaussian problem discussed
in the text, b) classification accuracy on Brodatz as a function of feature space dimension, and c)
corresponding curves of cumulative marginal density (9). A linear trend was subtracted to all curves
in c) to make the differences more visible.
that, in the context of visual processing, (8) will be approximately true and the MMD
principle will consequently lead to solutions that are very close to optimal, in the minimum
BE sense. Given the simplicity of MMD feature selection, this is quite remarkable. Second,
it implies that when combined with such transformations, the marginal diversity is a close
predictor for the CPE (and consequently the BE) achievable in a given feature space. This
enables quantifying the goodness of the transformation without even having to build the
classifier. See [13] for a more extensive discussion of these issues.
5 Experimental results
In this section we present results showing that 1) MMD feature selection outperforms combinatorial search when (8) holds, and 2) in the context of visual recognition problems,
marginal diversity is a good predictor of PE. We start by reporting results on a synthetic
problem, introduced by Trunk to illustrate the curse of dimensionality [12], and used by
Jain and Zongker (JZ) to evaluate various feature selection procedures
d : It consists of
d d ,.,-, [6].
and is an intwo Gaussian classes of identity covariance and means 7R) ?
?
teresting benchmark for feature selection because it has a clear optimal solution for the
best subset of [ features (the first [ ) for any [ . JZ exploited this property to propose
an automated procedure for testing the performance of feature selection algorithms across
variations in dimensionality of the feature space and sample size. We repeated their experiments, simply replacing the cost function they used (Mahalanobis distance - MDist between the means) by the marginal diversity.
Figure 4 a) presents the JZ score obtained with MMD as a function of the sample size. A
comparison with Figure 5 of [6] shows that these results are superior to all those obtained
by JZ, including the ones relying on branch and bound. This is remarkable, since branch
and bound is guaranteed to find the optimal solution and the Mdist is inversely proportional
to the PE for Gaussian classes. We believe that the superiority of MMD is due to the
fact that it only requires estimates of the marginals, while the MDist requires estimates
of joint densities and is therefore much more susceptible to the curse of dimensionality.
Unfortunately, because in [6] all results are averaged over dimension, we have not been
able to prove this conjecture yet. In any case, this problem is a good example of situations
where, because (8) holds, MMD will find the optimal solution at a computational cost that
is various orders of magnitude smaller than standard procedures based on combinatorial
search (e.g. branch and bound).
Figures 4 b) and c) show that, for problems involving commonly used image transformations, marginal diversity is indeed a good predictor of classification accuracy. The figures
compare, for each space dimension, the recognition accuracy of a complete texture recognition system with the predictions provided by marginal diversity. Recognition accuracy
was measured on the Brodatz texture database ( )) m texture classes) and a dimensional
feature space consisting of the coefficients of a multiresolution decomposition over regions
of pixels. Three transformations were considered: the discrete cosine transform, principal component analysis, and a three-level wavelet decomposition (see [14] for detailed
description of the experimental set up). The classifier was based on Gauss mixtures and
marginal diversity was computed with Algorithm 1. Note that the curves of cumulative
marginal diversity are qualitatively very similar to those of recognition accuracy.
References
[1] S. Basu, C. Micchelli, and P. Olsen. Maximum Entropy and Maximum Likelihood Criteria for
Feature Selection from Multivariate Data. In Proc. IEEE International Symposium on Circuits
and Systems, Geneva, Switzerland,2000.
[2] A. Bell and T. Sejnowski. An Information Maximisation Approach to Blind Separation and
Blind Deconvolution. Neural Computation, 7(6):1129?1159, 1995.
[3] B. Bonnlander and A. Weigand. Selecting Input Variables using Mutual Information and Nonparametric Density Estimation. In Proc. IEEE International ICSC Symposium on Artificial
Neural Networks, Tainan,Taiwan,1994.
[4] D. Erdogmus and J. Principe. Information Transfer Through Classifiers and its Relation to
Probability of Error. In Proc. of the International Joint Conference on Neural Networks, Washington, 2001.
[5] J. Huang and D. Mumford. Statistics of Natural Images and Models. In IEEE Computer Society
Conference on Computer Vision and Pattern Recognition, Fort Collins, Colorado, 1999.
[6] A. Jain and D. Zongker. Feature Selection: Evaluation, Application, and Small Sample Performance. IEEE Trans. on Pattern Analysis and Machine Intelligence, 19(2):153?158, February
1997.
[7] R. Linsker. Self-Organization in a Perceptual Network. IEEE Computer, 21(3):105?117, March
1988.
[8] J. Portilla and E. Simoncelli. Texture Modeling and Synthesis using Joint Statistics of Complex
Wavelet Coefficients. In IEEE Workshop on Statistical and Computational Theories of Vision,
Fort Collins, Colorado, 1999.
[9] J. Principe, D. Xu, and J. Fisher. Information-Theoretic Learning. In S. Haykin, editor, Unsupervised Adaptive Filtering, Volume 1: Blind-Souurce Separation. Wiley, 2000.
[10] G. Saon and M. Padmanabhan. Minimum Bayes Error Feature Selection for Continuous Speech
Recognition. In Proc. Neural Information Proc. Systems, Denver, USA, 2000.
[11] K. Torkolla and W. Campbell. Mutual Information in Learning Feature Transforms. In Proc.
International Conference on Machine Learning, Stanford, USA, 2000.
[12] G. Trunk. A Problem of Dimensionality: a Simple Example. IEEE Trans. on Pattern. Analysis
and Machine Intelligence, 1(3):306?307, July 1979.
[13] N. Vasconcelos. Feature Selection by Maximum Marginal Diversity: Optimality and Implications for Visual Recognition. In submitted, 2002.
[14] N. Vasconcelos and G. Carneiro. What is the Role of Independence for Visual Regognition? In
Proc. European Conference on Computer Vision, Copenhagen, Denmark, 2002.
[15] H. Yang and J. Moody. Data Visualization and Feature Selection: New Algorithms for Nongaussian Data. In Proc. Neural Information Proc. Systems, Denver, USA, 2000.
| 2169 |@word cpe:7 illustrating:1 middle:3 achievable:1 seek:1 covariance:1 decomposition:2 initial:1 contains:1 score:3 selecting:1 past:1 outperforms:1 current:3 yet:1 dct:2 additive:1 distant:1 visible:1 subsequent:1 enables:1 plot:5 discrimination:2 alone:1 v:1 intelligence:2 haykin:1 characterization:3 coarse:1 location:1 simpler:1 along:2 direct:2 symposium:2 consists:1 prove:1 indeed:2 p1:2 multi:1 relying:1 decreasing:2 little:2 curse:2 increasing:1 provided:1 bounded:2 linearity:1 maximizes:2 medium:1 mass:1 circuit:1 what:1 minimizes:3 extremum:2 transformation:10 finding:1 guarantee:2 concave:1 xd:1 shed:1 decouples:1 classifier:4 uk:1 appear:1 superiority:1 engineering:1 merge:1 approximately:1 suggests:2 co:1 averaged:1 testing:1 practice:1 maximisation:1 implement:1 procedure:8 universal:1 bell:1 significantly:1 projection:2 suggest:1 close:3 selection:32 operator:1 layered:1 context:7 impossible:1 pzq:5 optimize:1 equivalent:3 straightforward:3 simplicity:5 assigns:1 insight:2 rule:1 searching:1 variation:1 diego:1 play:2 colorado:2 jif:1 element:1 recognition:12 particularly:1 located:1 updating:1 trend:1 database:1 role:3 electrical:1 solved:1 region:1 decrease:1 removed:1 intuition:3 complexity:1 depend:1 tight:4 division:1 easily:1 joint:3 various:6 carneiro:1 derivation:2 separated:1 jain:3 sejnowski:1 artificial:1 quite:3 heuristic:1 stanford:1 plausible:2 favor:1 statistic:4 transform:4 advantage:2 sequence:1 propose:1 relevant:2 multiresolution:1 intuitive:1 description:1 zp:1 brodatz:2 derive:1 illustrate:1 measured:1 advocated:1 p2:2 implemented:1 implies:2 quantify:1 differ:1 concentrate:1 switzerland:1 closely:2 elimination:1 require:1 really:1 hold:5 considered:1 cb:1 major:1 vary:1 omitted:1 estimation:1 proc:9 combinatorial:5 label:6 expose:1 minimization:4 mit:1 clearly:2 gaussian:6 improvement:1 prevalent:1 likelihood:1 sense:14 typically:2 relation:1 transformed:1 pixel:1 issue:1 uncovering:1 classification:20 among:2 constrained:1 mutual:6 marginal:24 equal:3 once:1 vasconcelos:3 extraction:7 having:1 washington:1 identical:1 unsupervised:1 nearly:1 linsker:2 few:1 preserve:1 recognize:1 divergence:1 consisting:1 recalling:1 organization:2 righthand:1 evaluation:1 mixture:1 extreme:2 light:2 implication:2 lh:10 theoretical:1 mk:1 modeling:1 goodness:2 maximization:1 cost:7 subset:11 predictor:3 reported:1 synthetic:1 combined:2 density:9 international:4 discriminating:2 infomax:31 synthesis:1 quickly:1 moody:1 nongaussian:1 central:1 nm:2 successively:1 containing:1 huang:1 leading:1 return:2 potential:1 diversity:21 gaussianity:1 coefficient:2 blind:4 depends:1 performed:2 icsc:1 start:3 bayes:8 complicated:1 contribution:2 minimize:1 accuracy:5 weak:1 zy:2 emphasizes:1 submitted:1 reach:1 fo:1 definition:1 energy:1 nuno:2 obvious:1 associated:1 mi:2 proof:2 gain:1 knowledge:1 dimensionality:4 campbell:1 follow:1 reflected:1 maximally:1 bonnlander:1 though:2 furthermore:4 hand:3 replacing:1 believe:1 usa:3 concept:1 true:1 contain:1 former:1 equality:1 inspiration:1 leibler:1 illustrated:1 attractive:1 mahalanobis:1 self:1 cosine:1 criterion:7 theoretic:3 complete:1 fj:2 image:8 wise:1 consideration:1 recently:2 superior:1 denver:2 foreach:2 volume:1 discussed:1 marginals:4 significant:2 fano:4 zongker:3 surface:2 posterior:1 multivariate:1 recent:3 optimizing:1 inequality:1 exploited:1 seen:1 minimum:7 recognized:1 maximize:2 july:1 branch:4 d7:4 simoncelli:1 reduces:1 long:1 prediction:1 involving:2 regression:1 vision:3 expectation:2 histogram:1 mmd:17 achieved:1 preserved:1 source:1 standpoint:2 unlike:2 ee:1 near:4 noting:2 yang:1 recommends:2 automated:1 independence:1 identified:1 opposite:1 regarding:2 expression:1 pca:2 ultimate:1 speech:2 proceed:1 useful:2 clear:6 involve:1 detailed:2 amount:2 transforms:2 nonparametric:1 processed:1 diverse:1 discrete:1 affected:1 key:1 drawn:6 backward:2 year:1 reporting:1 family:3 throughout:1 separation:4 decision:1 layer:2 bound:22 guaranteed:2 constraint:1 x2:1 optimality:5 px:3 conjecture:2 department:1 according:1 march:1 saon:1 across:1 smaller:1 appealing:1 biologically:2 making:1 ikj:2 equation:4 visualization:2 trunk:2 turn:1 serf:1 available:2 tightest:1 generic:3 subtracted:1 alternative:2 existence:1 denotes:1 top:1 build:1 establish:2 february:1 classical:1 society:1 micchelli:1 question:1 quantity:3 contacting:1 added:1 strategy:2 mumford:1 dependence:3 prq:6 said:1 distance:2 separate:1 sensible:1 discriminant:6 reason:2 denmark:1 taiwan:1 assuming:1 besides:2 relationship:5 minimizing:4 difficult:1 susceptible:1 unfortunately:1 observation:3 benchmark:1 padmanabhan:1 defining:1 situation:1 precise:1 portilla:1 introduced:3 fort:2 copenhagen:1 kl:1 extensive:1 connection:3 california:1 trans:2 address:1 able:1 pattern:5 summarize:1 encompasses:1 max:1 including:1 unrealistic:1 power:1 natural:2 examination:1 rely:1 indicator:2 mdist:3 inversely:1 lk:3 axis:1 text:1 prior:3 literature:1 law:1 interesting:4 limitation:1 proportional:1 filtering:1 versus:2 vg:1 remarkable:2 conveyed:1 consistent:1 principle:18 editor:1 side:1 basu:1 curve:3 dimension:5 cumulative:3 contour:3 author:2 forward:2 commonly:1 san:1 qualitatively:1 adaptive:1 geneva:1 ignore:1 olsen:1 kullback:1 ml:1 instantiation:1 reveals:1 weigand:1 recommending:1 search:9 continuous:1 why:1 nature:1 jz:5 transfer:1 inherently:1 complex:2 european:1 main:3 rh:9 s2:1 n2:1 repeated:1 x1:1 xu:1 wiley:1 perceptual:2 pe:9 wavelet:4 theorem:8 specific:1 showing:1 r2:1 evidence:1 deconvolution:1 intrinsic:1 workshop:1 sequential:2 adding:1 texture:4 magnitude:1 easier:1 entropy:5 simply:1 visual:8 satisfies:1 conditional:2 goal:3 formulated:1 identity:1 consequently:3 quantifying:1 erdogmus:1 fisher:1 considerable:1 change:1 determined:1 lemma:2 principal:2 called:1 experimental:3 gauss:1 select:2 principe:2 support:1 latter:2 collins:2 evaluate:1 ex:1 |
1,284 | 217 | VLSI Implementation of a High-Capacity Neural Network
VLSI Implementation of a High-Capacity
Neural Network Associative Memory
Tzi-Dar Chiueh 1 and Rodney M. Goodman
Department of Electrical Engineering (116-81)
California Institute of Technology
Pasadena, CA 91125, USA
ABSTRACT
In this paper we describe the VLSI design and testing of a high
capacity associative memory which we call the exponential correlation associative memory (ECAM). The prototype 3J.'-CMOS
programmable chip is capable of storing 32 memory patterns of
24 bits each. The high capacity of the ECAM is partly due to the
use of special exponentiation neurons, which are implemented via
sub-threshold MOS transistors in this design. The prototype chip
is capable of performing one associative recall in 3 J.'S.
1
ARCHITECTURE
Previously (Chiueh, 1989), we have proposed a general model for correlation-based
associative memories, which includes a variant of the Hopfield memory and highorder correlation memories as special cases. This new exponential correlation associative memory (ECAM) possesses a very large storage capacity, which scales
exponentially with the length of memory patterns (Chiueh, 1988). Furthermore, it
has been shown that the ECAM is asymptotically stable in both synchronous and
1Tzi-Dar Chiueh is now with the Department of Electrical Engineering, National Taiwan University, Taipei, Taiwan 10764.
793
794
Chiueh and Goodman
asynchronous updating modes (Chiueh, 1989). The model is based on an architecture consisting of binary connection weights, simple hard-limiter neurons, and
specialized nonlinear circuits as shown in Figure 1. The evolution equation of this
general model is
(1)
where u(1), u(2), ... , u(M) are the M memory patterns. x and x, are the current and
the next state patterns of the system respectively, and sgn is the threshold function,
which takes on the value +1 if its argument is nonnegative, and -1 otherwise.
We addressed, in particular, the case where f(?) is in the form of an exponentiation,
namely, when the evolution equation is given by
(2)
and a is a constant greater than unity.
The ECAM chip we have designed is programmable; that is, one can change the
stored memory patterns at will. To perform an associative recall, one first loads a
set of memory patterns into the chip. The chip is then switched to the associative
recall mode, an input pattern is presented to the ECAM chip, and the ECAM chip
then computes the next state pattern according to Equation (2). The components
of the next state pattern appear at the output in parallel after the internal circuits
have settled. Feedback is easily incorporated by connecting the output port to the
input port, in which case the chip will cycle until a fixed point is reached.
2
DESIGN OF THE ECAM CIRCUITS
From the evolution equation of the ECAM, we notice that there are essentially three
circuits that need to be designed in order to build an ECAM chip. They are:
? < u(1:), x >, the
correlation computation circuit;
M
? I:
a<u(k), x>
u(1:),
the exponentiation, multiplication and summing circuit;
1:=1
? sgn( .), the threshold circuit.
We now describe each circuit, present its design, and finally integrate all these
circuits to get the complete design of the ECAM chip.
VLSI Implementation ora High-Capacity Neural Network
2.1
CORRELATION COMPUTATION
In Figure 2, we illustrate a voltage-divider type circuit consisting of NMOS transistors working as controlled resistors (linear resistors or open circuits). This circuit
computes the correlation between the input pattern x and a memory pattern u(l:).
If the ith components of these two patterns are the same, the corresponding XOR
gate outputs a "0" and there is a connection from the node V~~ to VBB; otherwise, there is a connection from V~~ to GND. Hence the output voltage will be
proportional to the number of positions at which x and u(l:) match. The maximum
output voltage is controlled by an externally supplied bias voltage VBB. Normally,
VBB is set to a voltage lower than the threshold voltage of NMOS transistors (VTH)
for a reason that will be explained later. Note that the conductance of an NMOS
transistor in the ON mode is not fixed, but rather depends on its gate-to-source
voltage and its drain-to-source voltage. Thus, some nonlinearity is bound to occur
in the correlation computation circuit, however, simulation shows that this effect is
small.
2.2
EXPONENTIATION, MULTIPLICATION, AND SUMMATION
Figure 4 shows a circuit that computes the exponentiation of V~~, the product of
the u~l:) and the exponential, and the sum of all M products.
The exponentiation function is implemented by an NMOS transistor whose gate
voltage is V~~. Since VBB, the maximum value that V~~ can assume, is set to be
lower than the threshold voltage (VTH); the NMOS transistor is in the subthreshold
region, where its drain current depends exponentially on its gate-to-source voltage
(Mead, 1989). If we temporarily ignore the transistors controlled by u~l:) or the
complement of u~l:), the current flowing through the exponentiation transistor associated with V~~ will scale exponentially with V~~. Therefore, the exponentiation
function is properly computed.
Since the multiplier u~l:) assumes either +1 or -1, the multiplication can be easily
done by forming two branches, each made up of a transmission gate in series with an
exponentiation transistor whose gate voltage is V~~. One of the two transmission
gates is controlled by u~l:), and the other by the complement of u~l:). Consequently,
when u~l:)
1, the positive branch will carry a current that scales exponentially
with the correlation of the input x and the ph memory pattern u(l:) , while the
negative branch is essentially an open circuit, and vice versa.
=
Summation of the M terms in the evolution equation is done by current summing.
The final results are two currents It and Ii, which need to be compared by a
threshold circuit to determine the sign of the ith bit of the next state pattern x~.
In the ECAM a simple differential amplifier (Figure 3) performs the comparison.
795
796
Cbiueb and Goodman
2.3
THE BASIC ECAM CELL
The above computational circuits are then combined with a simple static RAM cell,
to make up a basic ECAM cell as illustrated in Figure 5. The final design of an
ECAM that stores M N-bit memory patterns can be obtained by replicating the
basic ECAM cell M times in the horizontal direction and N times in the vertical
direction, together with read/write circuits, sense amplifiers, address decoders, and
I/O multiplexers. The prototype ECAM chip is made up of 32 x 24 ECAM cells,
and stores 32 memory patterns each 24 bits wide.
3
ECAM CHIP TEST RESULTS
The test procedure for the ECAM is to first generate 32 memory patterns at random
and then program the ECAM chip with these 32 patterns. We then pick a memory
pattern at random, flip a specified number of bits randomly, and feed the resulting
pattern to the ECAM as an input pattern (x). The output pattern (x') can then be
fed back to the inputs of the ECAM chip. This iteration continues until the pattern
at the input is the same as that at the the output, at which time the ECAM chip
is said to have reached a stable state. We select 10 sets of 32 memory patterns and
for each set we run the ECAM chip on 100 trial input patterns with a fixed number
of errors. Altogether, the test consists of 1000 trials.
In Figure 6, we illustrate the ECAM chip test results. The number of successes is
plotted against the number of errors in the input patterns for the following four
cases: 1) The ECAM chip with VBB
5V; 2) VBB
2V; 3) VBB
IV; and 4) a
simulated ECAM in which the exponentiation constant a, equals 2. It is apparent
from Figure 6 that as the number of errors increases, the number of successes
decreases, which is expected. Also, one notices that the simulated ECAM is by far
the best one, which is again not unforeseen because the ECAM chip is, after all,
only an approximation of the ideal ECAM model.
=
=
=
What is really unexpected is that the best performance occurs for VBB = 2V rather
than VBB = IV (VTH in this CMOS process). This phenomenon arises because
of two contradictory effects brought about by increasing VBB. On the one hand,
increasing VBB increases the dynamic range of the exponentiation transistors in the
ECAM chip. Suppose that the correlations of two memory patterns u(l) and u(k)
with the input pattern x are tJ and tk, respectively, where tJ > tk; then
V(I) _ (tJ
ux -
+ N) VBB
2N
(k) _
-
,V ux
(tk
+ N)
2N
VBB
.
Therefore, as VBB increases, so does the difference between V~I~ and V~~, and u(l)
becomes more dominant than u(k) in the weighted sum of the evolution equation.
VLSI Implementation or a High?Capacity Neural Network
Hence, as VBB increases, the error correcting ability of the ECAM chip should
improve. On the other hand, as VBB increases beyond the threshold voltage, the
exponentiation transistors leave the subthreshold region and may enter saturation,
where the drain current is approximately proportional to the square of the gateto-source voltage . Since a second-order correlation associative memory in general
possesses a smaller storage capacity than an ECAM, one would expect that with
a fixed number of loaded memory patterns, the ECAM should do better than the
second-order correlation associative memory. Thus one effect tends to enhance the
performance of the ECAM chip, while the other tends to degrade it. A compromise
between these two effects is reached, and the best performance is achieved when
VBB = 2V.
=
For the case when VBB 2V, the drain current versus gate-to-source voltage characteristic of the exponentiation transistors is actually a hybrid of a square function
and an exponentiation function. At the bottom it is of an exponential form, and
it gradually flattens out to a square function, once the gate-to-source voltage becomes larger than the threshold voltage . Therefore, the ECAM chip with VBB
2V is a mixture of the second-order correlation associative memory and the pure
ECAM . According to the convergence theorem for correlation associative memories
(Chiueh, 1989) and the fact that f(?) in the ECAM chip with VBB
2V is still
monotonically nondecreasing, the ECAM chip is still asymptotically stable when
VBB
2V.
=
=
=
We have tested the speed of the ECAM chip using binary image vector quantization
as an example problem. The speed at which the ECAM chip can vector-quantize
binary images is of interest. We find experimentally that the ECAM chip is capable
of doing one associative recall operation, in less than 3 j.ts, ' n 4 x 4 blocks. This
projects to approximately 49 ms for a 512 x 512 binary image, or more than 20
images per second .
4
CONCLUSIONS
In this paper, we have presented a VLSI circuit design for implementing a high
capacity correlation associative memory. The performance of the ECAM chip is
shown to be almost as good as a computer-simulated ECAM . Furthermore, we
believe that the ECAM chip is more robust than an associative memory using a
winner-take-all function, because it obtains its result via iteration, as opposed to
one shot. In conclusion, we believe that the ECAM chip provides a fast and efficient
way for solving many associative recall problems, such as vector quantization and
optical character recognition.
Acknowledgement
This work was supported in part by NSF grant No . MIP - 8711568.
797
798
Chiueh and Goodman
References
T. D. Chiueh and R. M. Goodman. (1988) "High Capacity Exponential Associative
Memory," in Proc. of IEEE IeNN, Vol. I, pp. 153-160.
T. D. Chiueh. (1989) "Pattern Classification and Associative Recall by Neural
Networks," Ph. D. dissertation, California Institute of Technology.
C. A. Mead. (1989) Analog VLSI and Neural Systems. Reading, MA : AddisonWesley.
Figure 1: Architecture of the General Correlation-Based Associative Memory
(I<)
u
N-l
X
N-l
Figure 2: The Correlation Computation Circuit
VLSI Implementation or a High-Capacity Neural Network
Voo
X'.
I
I
+
I
Figure 3: The Threshold Circuit
-
Ii
V (1)
ux
(1)
U.
I
V
(2)
ux
(2)
U.
I
?
??
V(M)
ux
(M)
U.
I
Figure 4: The Exponentiation, Multiplication, and Summation Circuit
799
800
Chiueh and Goodman
I 1.
I 1.
r.,
(1<)
U.
1
V(k)
ux
(1<)
RAM
u.1
cell
(1<)
u.1
Figure 5: Circuit Diagram of the Basic ECAM Cell
.....res
1000
G
?
fIl
....
900
8.....
.....
800
.~
0
....0
g
~fIl
~
~
.....0
'"'
Q)
]
Z
700
600
.. Simulation (a=2)
500
... Vbb=5V
400
300
0-
Vbb =2V
-I-
Vbb = lV
200
100
0
0
1
2
3
4
5
6
7
Number of errors in input patterns
Figure 6: Error Correcting Ability of the ECAM Chip with Different
with a Simulated ECAM with a 2
=
VBB
compared
| 217 |@word effect:4 trial:2 implemented:2 divider:1 multiplier:1 build:1 evolution:5 hence:2 direction:2 read:1 open:2 flattens:1 occurs:1 simulation:2 illustrated:1 sgn:2 said:1 pick:1 implementing:1 shot:1 simulated:4 carry:1 capacity:11 m:1 decoder:1 series:1 really:1 degrade:1 complete:1 summation:3 reason:1 performs:1 taiwan:2 current:8 fil:2 length:1 image:4 mo:1 specialized:1 multiplication:4 negative:1 winner:1 designed:2 exponentially:4 analog:1 proc:1 implementation:5 design:7 perform:1 limiter:1 vertical:1 neuron:2 versa:1 enter:1 ith:2 vice:1 dissertation:1 t:1 weighted:1 provides:1 brought:1 node:1 nonlinearity:1 replicating:1 incorporated:1 rather:2 stable:3 voltage:17 differential:1 dominant:1 complement:2 consists:1 namely:1 specified:1 properly:1 connection:3 store:2 california:2 expected:1 binary:4 os:1 sense:1 success:2 address:1 beyond:1 tested:1 greater:1 pattern:31 pasadena:1 vlsi:8 determine:1 reading:1 increasing:2 becomes:2 project:1 monotonically:1 branch:3 ii:2 circuit:23 classification:1 program:1 saturation:1 what:1 memory:29 hybrid:1 match:1 special:2 equal:1 once:1 improve:1 technology:2 controlled:4 variant:1 basic:4 essentially:2 iteration:2 tzi:2 normally:1 grant:1 achieved:1 appear:1 randomly:1 cell:7 acknowledgement:1 positive:1 national:1 engineering:2 addressed:1 drain:4 tends:2 diagram:1 source:6 consisting:2 expect:1 goodman:6 proportional:2 mead:2 posse:2 amplifier:2 conductance:1 versus:1 approximately:2 interest:1 lv:1 switched:1 integrate:1 chiueh:11 call:1 mixture:1 port:2 storing:1 ideal:1 range:1 tj:3 supported:1 testing:1 architecture:3 capable:3 block:1 asynchronous:1 bias:1 prototype:3 institute:2 procedure:1 addisonwesley:1 iv:2 wide:1 synchronous:1 re:1 plotted:1 mip:1 altogether:1 feedback:1 computes:3 made:2 get:1 far:1 storage:2 dar:2 programmable:2 select:1 obtains:1 ignore:1 ph:2 gnd:1 stored:1 summing:2 generate:1 supplied:1 correcting:2 pure:1 nsf:1 combined:1 notice:2 sign:1 per:1 robust:1 write:1 ca:1 enhance:1 connecting:1 together:1 unforeseen:1 vol:1 suppose:1 four:1 again:1 threshold:9 settled:1 quantize:1 opposed:1 recognition:1 updating:1 continues:1 ram:2 asymptotically:2 multiplexer:1 sum:2 bottom:1 run:1 exponentiation:15 electrical:2 includes:1 region:2 almost:1 cycle:1 depends:2 sub:1 decrease:1 later:1 position:1 resistor:2 exponential:5 doing:1 bit:5 reached:3 bound:1 parallel:1 dynamic:1 rodney:1 externally:1 highorder:1 nonnegative:1 theorem:1 solving:1 square:3 occur:1 compromise:1 xor:1 loaded:1 characteristic:1 load:1 subthreshold:2 easily:2 hopfield:1 chip:32 speed:2 argument:1 nmos:5 quantization:2 performing:1 optical:1 fast:1 describe:2 department:2 voo:1 according:2 smaller:1 whose:2 apparent:1 larger:1 against:1 character:1 unity:1 pp:1 otherwise:2 forming:1 vth:3 ability:2 unexpected:1 associated:1 explained:1 gradually:1 static:1 nondecreasing:1 temporarily:1 ux:6 final:2 associative:19 equation:6 recall:6 transistor:12 previously:1 ma:1 consequently:1 product:2 flip:1 actually:1 back:1 fed:1 feed:1 hard:1 change:1 experimentally:1 operation:1 flowing:1 contradictory:1 done:2 furthermore:2 partly:1 correlation:17 convergence:1 until:2 transmission:2 working:1 horizontal:1 hand:2 gate:9 cmos:2 leave:1 nonlinear:1 tk:3 assumes:1 illustrate:2 internal:1 mode:3 arises:1 believe:2 taipei:1 phenomenon:1 usa:1 |
1,285 | 2,170 | Retinal Processing Emulation in a
Programmable 2-Layer Analog Array
Processor CMOS Chip
R. Carmona, F. Jim?
enez-Garrido, R. Dom??nguez-Castro,
S. Espejo, A. Rodr??guez-V?
azquez
Instituto de Microelectr?
onica de Sevilla-CNM-CSIC
Avda. Reina Mercedes s/n 41012 Sevilla (SPAIN)
[email protected]
Abstract
A bio-inspired model for an analog programmable array processor
(APAP), based on studies on the vertebrate retina, has permitted
the realization of complex programmable spatio-temporal dynamics in VLSI. This model mimics the way in which images are processed in the visual pathway, rendering a feasible alternative for
the implementation of early vision applications in standard technologies. A prototype chip has been designed and fabricated in a
0.5?m standard CMOS process. Computing power per area and
power consumption is amongst the highest reported for a single
chip. Design challenges, trade-offs and some experimental results
are presented in this paper.
1
Introduction
The conventional role of analog circuits in mixed-signal VLSI is providing the I/O
interface to the digital core of the chip ?which realizes all the signal processing.
However, this approach may not be optimum for the processing of multi-dimensional
sensory signals, such as those found in vision applications. When massive information flows have to be treated in parallel, it may be advantageous to realize some
preprocessing in the analog domain, at the plane where signals are captured.
During the last years, different authors have focused on the realization of parallel
preprocessing of multi-dimensional signals, using either purely digital techniques [1]
or mixed-signal techniques, like in [2]. The data in Table 1 can help us to compare
these two approaches. Here, the peak computing power (expressed as operations
per second: XPS) per unit area and power is shown. This estimation is realized
by considering the number of arithmetic analog operations that take place per unit
time, in the analog case, or digital instructions per unit time, in the digital case. It
can be seen that the computing power per area featured by chips based in Analog
Programmable Array Processors (APAPs) is much higher than that exhibited by
digital array processors. It can be argued that digital processors feature a larger
accuracy, but accuracy requirements for vision applications are not rarely below 6
Table 1: Parallel processors comparison
Reference
Li?
nan et. al. [2]
Gealow et. al. [1]
This chip
CMOS
process
No.
of cells
Cells/
mm2
XPS/
mm2
XPS/
mW
0.5?m
0.6?m
0.5?m
4096
4096
1024
81.0
66.7
29.2
7.93G
4.00M
6.01G
0.33G
1.00M
1.56G
bits. Also, taking full advantage of the full digital resolution requires highly accurate
A/D converters, what creates additional area and power overhead.
The third row in Table 1 corresponds to the chip presented here. This chip outperforms the one in [2] in terms of functionality as it implements a reduced model
of the biological retina [3]. It is capable of generating complex spatio-temporal dynamic processes, in a fully programmable way and with the possibility of storing
intermediate processing results.
2
2.1
APAP chip architecture
Bio-inspired APAP model
The vertebrate retina has a layered structure [3], consisting, roughly, in a layer of
photodetectors at the top, bipolar cells carrying signals across the retina, affected by
the operation of horizontal and amacrine cells, and ganglion cells in the other end.
There are, in this description, some interesting aspects that markedly resemble
the characteristics of the Cellular Neural Networks (CNNs) [4]: 2D aggregations
of continuous signals, local connectivity between elementary nonlinear processors,
analog weighted interactions between them. Motivated by these coincidences, a
model consisting of 2 layers of processors coupled by some inter-layer weights, and an
additional layer incorporating analog arithmetics, has been developed [5]. Complex
dynamics can be programmed via the intra- and inter-layer coupling strengths and
the relation between the time constants of the layers. The evolution of each cell,
C(i, j), is described by two coupled differential equations, one for each CNN node:
?n
r1
X
dxn,ij
= ?g[xn,ij (t)] +
dt
r1
X
ann,kl ? yn,(i+k)(j+l) +
k=?r1 l=?r1
+bnn,00 ? unn,ij + zn,ij + ano ? yno,ij
(1)
where n and o stand for the node in question and the other node respectively. The
nonlinear losses term and the output function in each layer are those described for
the full-signal range (FSR) model of the CNN [7], in which the state voltage is also
limited and can be identified with the output voltage:
g(xn,ij ) = lim
m??
and:
(
m(xn,ij ? 1) + 1
xn,ij
?m(xn,ij + 1) ? 1
if
xn,ij > 1
if |xn,ij | ? 1
if xn,ij < ?1
(2)
yn,ij = f (xn,ij ) =
1
(|xn,ij + 1| ? |xn,ij ? 1|)
2
(3)
The proposed chip consists in an APAP of 32 ? 32 identical 2nd-order CNN cells
(Fig. 3), surrounded by the circuits implementing the boundary conditions.
Figure 1: (a) Conceptual diagram of the basic cell and (b) internal structure of each
CNN layer node
2.2
Basic processing cell architecture
Each elementary processor includes two coupled continuous-time CNN cores
(Fig. 1(a)). The synaptic connections between processing elements of the same
or different layer are represented by arrows in the diagram. The basic processor
contains also a programmable local logic unit (LLU) and local analog and logic
memories (LAMs and LLMs) to store intermediate results. The blocks in the cell
communicate via an intra-cell data bus, multiplexed to the array interface. Control
bits and switch configuration are passed to the cell from a global programming unit.
The internal structure of each CNN core is depicted in the diagram of Fig. 1(b).
Each core receives contributions from the rest of the processing nodes in the neighbourhood which are summed and integrated in the state capacitor. The two layers
differ in that the first layer has a scalable time constant, controlled by the appropriate binary code, while the second layer has a fixed time constant. The evolution
of the state variable is also driven by self-feedback and by the feedforward action of
the stored input and bias patterns. There is a voltage limiter for implementing the
FSR CNN model. Forcing the state voltage to remain between these limits allows
for using it as the output voltage. Then the state variable, which is now the output, is transmitted in voltage form to the synaptic blocks, in the periphery of the
cell, where weighted contributions to the neighbours? are generated. There is also a
current memory that will be employed for cancellation of the offset of the synaptic
blocks. Initialization of the state, input and/or bias voltages is done through a
mesh of multiplexing analog switches that connect to the cell?s internal data bus.
3
3.1
Analog building blocks for the basic cell
Single-transistor synapse
The synapse is a four-quadrant analog multiplier. Their inputs will be the cell
state, or input, and the weight voltages, while the output will be the cell?s current
contribution to a neighbouring cell. It can be realized by a single transistor biased
in the ohmic region [6]. For a PMOS with gate voltage VX = Vx0 + Vx , and the
p-diffusion terminals at VW = Vw0 + Vw and Vw , the drain-to-source current is:
Vw
Io ? ??p Vw Vx ? ?p Vw Vx0 + |V?Tp | ? Vw0 ?
2
(4)
which is a four-quadrant multiplier with an offset term that is time-invariant ?at
least during the evolution of the network? and not depending on the state. This
offset is eliminated in a calibration step, with a current memory.
For the synapse to operate properly, the input node of the CNN core,
L in Fig. 2,
must be kept at a constant voltage. This is achieved by a current conveyor. Any
difference between the voltage at node
L and the reference V w0 is amplified and
the negative feedback corrects the deviation. Notice that a voltage offset in the
amplifier results in an error of the same order. An offset cancellation mechanism is
provided (Fig. 2).
3.2
S3 I current memory
As it has been referred, the offset term of the synapse current must be removed
for its output current to represent the result of a four-quadrant multiplication. For
this purpose all the synapses are reset to VX = Vxo . Then the resulting current,
which is the sum of the offset currents of all the synapses concurrently connected
to the same node, is memorized. This value will be substracted on-line from the
input current when the CNN loop is closed, resulting in a one-step cancellation of
the errors of all the synapses. The validity of this method relies in the accuracy of
the current memory. For instance, in this chip, the sum of all the contributions will
range, for the applications for which it has been designed, from 18?A to 46?A. On
the other side, the maximum signal to be handled is 1?A. If a signal resolution of
8b is pretended, then 0.5LSB = 2nA. Thus, our current memory must be able to
distinguish 2nA out of 46?A. This represents an equivalent resolution of 14.5b. In
order to achieve such accuracy level, a S3 I current memory is used. It is composed by
three stages (Fig. 2), each one consisting in a switch, a capacitor and a transistor. I B
is the current to be memorized. After memorization the only error left corresponds
to the last stage.
3.3
Time-constant scaling
The differential equation that governs the evolution of the network (1) can be
written as a sum of current contributions injected to the state capacitor. Scaling up/down this sum of currents is equivalent to scaling the capacitor and, thus,
speeding up/down the network dynamics. Therefore, scaling the input current with
the help of a current mirror, for instance, will have the effect of scaling the timeconstant. A circuit for continuously adjusting the current gain of a mirror can be
designed based on a regulated-Cascode current mirror in the ohmic region. But the
strong dependence of the ohmic-region biased transistors on the power rail voltage
causes mismatches in ? between cells in the same layer. An alternative to this is
a digitally programmable current mirror. It trades resolution in ? for robustness,
hence, the mismatch between the time constants of the different cells is now fairly
attenuated.
Figure 2: Input block with current scaling, S3 I memory and offset-corrected OTA
schematic
A new problem arises, though, because of current scaling. If the input current
can be reshaped to a 16-times smaller waveform, then the current memory has to
operate over a larger dynamic range. But, if designed to operate on large currents,
the current memory will not work for the tiny currents of the scaled version of the
input. If it is designed to run on small input currents, long transistors will be needed,
and the operation will be unreliable for the larger currents. One way of avoiding
this situation is to make the S3 I memory to work on the original unscaled version
of the input current. Therefore, the adjustable-time-constant CNN core will be a
current conveyor, followed by the S3 I current memory and then the binary weighted
current mirror. The problem now is that the offsets introduced by the scaling block
add up to the signal and the required accuracy levels can be lost. Our proposal is
depicted in Fig. 2. It consists in placing the scaling block (programmable mirror)
between the current conveyor and the current memory. In this way, any offset error
will be cancelled in the auto-zeroing phase. In the picture, the voltage reference
generated with the current conveyor, the regulated-Cascode current mirrors and
the S3 I memory can be easily identified. The inverter, Ai , driving the gates of the
transistors of the current memory is required for stability.
4
Chip data and experimental results
A prototype chip has been designed and fabricated in a 0.5?m single-poly triplemetal CMOS technology. Its dimensions are 9.27 ? 8.45mm2 (microphotograph in
Fig. 3). The cell density achieved is 29.24cells/mm2, once the overhead circuitry is
detracted from the total chip area ?given that it does not scale linearly with the
number of cells. The power consumption of the whole chip is around 300mW. Data
I/O rates are nominally 10MS/s. Equivalent resolution for the analog images handled by the chip is 7.5 bit (measured). The time constant of the fastest layer (fixed
time constant) is intended to be under 100ns. The peak computing power of this
chip is, therefore, 470GXPS, what means 6.01GXPS/mm2 , and 1.56GXPS/mW.
Figure 3: Prototype chip photograph
The programmable dynamics of the chip permit the observation of different phenomena of the type of propagation of active waves, pattern generation, etc. By
tuning the coefficients that control the interactions between the cells in the array?
i. e. the weights of the synaptic blocks, which are common to every elementary
processor? different dynamics are manifested. Fig. 4 displays the evolution of the
state variables of the two coupled layers when it is programmed to show different
propagative behaviors. In picture (a), the chip is programmed to resemble the socalled wide-field erasure effect observed in the retina. Markers in the fastest layer
(bottom row) trigger wavefronts in this layer and induce slower waves in the other
layer (upper row). These induced spots are fedback, inhibiting the waves propagating in the fast layer, and generating a trailing edge for each wavefront. In picture
(b), a solitary traveling wave is triggered from each corner of the fast layer. This
kind of behavior is proper of waves in active media. Finally, in picture (c), edge
detection is computed by extraction the low frequency components of the image,
obtained by a diffusion in the slower layer, from theoriginal one. The remaining
information is that of the higher frequency components of the image. These phenomena have been widely observed in measurements of the vertebrate retina [3].
They constitute the patterns of activity generated by the presence of visual stimuli.
Controlling the network dynamics and combining the results with the help of the
built-in local logic and arithmetic operators, rather involved image processing tasks
can be programmed like active-contour detection, object-tracking, etc.
5
Conclusions
From the figures obtained, we can state that the proposed approach supposes a
promising alternative to conventional digital image processing for applications re-
lated with early-vision and low-level focal-plane image processing. Based on a simple
but precise model of part of the real biological system, a feasible efficient implementation of an artificial vision device has been designed. The peak operation speed
of the chip outperforms its digital counterparts due to the fully parallel nature of
the processing. This especially so when comparing the computing power per silicon
area unit and per watt.
Acknowledgments
This work has been partially supported by ONR/NICOP Project N00014-00-1-0429,
ESPRIT V Project IST-1999-19007, and by the Spanish CICYT Project TIC-19990826.
References
[1] Gealow, J.C. & Sodini, C.G. (1999) A Pixel Parallel Image Processor Using
Logic Pitch -Matched to Dynamic Memory. IEEE Journal of Solid-State Circuits, Vol. 34, No. 6, pp. 831-839.
[2] Li?
nan, G., Espejo, S., Dom??nguez-Castro, R., Roca, E. and Rodr??guezV?
azquez, A. (1998) A 64 x 64 CNN with Analog and Digital I/O. Proceedings of the IEEE Int. Conf. on Electronics, Circuits and Systems, pp. 203-206,
Lisbon, Portugal.
[3] Werblin, F. (1991) Synaptic Connections, Receptive Fields and Patterns of Activity in the Tiger Salamander Retina. Investigative Ophthalmology and Visual
Science, Vol. 32, No. 3, pp. 459-483.
[4] Werblin, F., Roska, T. and Chua, L.O. (1995) The Analogic Cellular Neural
Network as a Bionic Eye. International Journal of Circuit Theory and Applications, Vol. 23, No. 6, pp. 541-69.
[5] Rekeczky, Cs., Serrano-Gotarredona, T., Roska, T. and Rodr??guez-V?
azquez, A.
(2000) A Stored Program 2nd Order/3- Layer Complex Cell CNN-UM. Proc.
of the Sixth IEEE International Workshop on Cellular Neural Networks and
their Applications, pp. 219-224, Catania, Italy.
[6] Dom??nguez-Castro, R., Rodr??guez-V?
azquez, A., Espejo, S. and Carmona, R.
(1998) Four-Quadrant One-Transistor Synapse for High Density CNN Implementations. Proc. of the Fifth IEEE International Workshop on Cellular Neural
Networks and their Applications, pp. 243-248, London, UK.
[7] Espejo, S., Carmona, R. Carmona, Dom??nguez-Castro, R. and Rodr??guezV?
azquez, A. (1996) A VLSI Oriented Continuous- Time CNN Model. International Journal of Circuits Theory and Applications, Vol. 24, No. 3, pp. 341-356,
John Wiley and Sons Ed.
Figure 4: Examples of the different dynamics that can be programmed on the chip:
(a) wide-field erasure effect, (b) traveling wave accross the layers, and (c) edge
detection.
| 2170 |@word cnn:14 version:2 advantageous:1 nd:2 instruction:1 solid:1 electronics:1 configuration:1 contains:1 outperforms:2 current:40 comparing:1 guez:3 must:3 written:1 john:1 realize:1 mesh:1 ota:1 designed:7 device:1 yno:1 plane:2 core:6 chua:1 propagative:1 node:8 differential:2 consists:2 pathway:1 overhead:2 inter:2 behavior:2 roughly:1 multi:2 terminal:1 inspired:2 accross:1 considering:1 vertebrate:3 spain:1 provided:1 project:3 matched:1 circuit:7 medium:1 what:2 tic:1 kind:1 developed:1 fabricated:2 temporal:2 every:1 bipolar:1 um:1 esprit:1 scaled:1 uk:1 lated:1 bio:2 unit:6 control:2 yn:2 werblin:2 local:4 limit:1 instituto:1 io:1 initialization:1 fastest:2 programmed:5 limited:1 range:3 acknowledgment:1 wavefront:2 lost:1 block:8 implement:1 spot:1 erasure:2 area:6 featured:1 induce:1 quadrant:4 layered:1 operator:1 memorization:1 conventional:2 equivalent:3 focused:1 resolution:5 array:6 stability:1 controlling:1 trigger:1 massive:1 programming:1 neighbouring:1 element:1 observed:2 role:1 bottom:1 coincidence:1 gealow:2 region:3 connected:1 trade:2 highest:1 removed:1 digitally:1 cnm:2 cascode:2 dynamic:10 dom:4 carrying:1 purely:1 creates:1 easily:1 chip:22 represented:1 ohmic:3 fast:2 investigative:1 london:1 artificial:1 pmos:1 larger:3 widely:1 reshaped:1 advantage:1 triggered:1 transistor:7 lam:1 interaction:2 reset:1 serrano:1 loop:1 realization:2 combining:1 achieve:1 amplified:1 description:1 optimum:1 r1:4 requirement:1 generating:2 cmos:4 object:1 help:3 coupling:1 depending:1 propagating:1 measured:1 ij:16 strong:1 c:1 resemble:2 differ:1 waveform:1 emulation:1 functionality:1 cnns:1 vx:4 memorized:2 implementing:2 microphotograph:1 argued:1 biological:2 elementary:3 around:1 circuitry:1 garrido:1 driving:1 inhibiting:1 trailing:1 early:2 inverter:1 purpose:1 estimation:1 proc:2 realizes:1 limiter:1 weighted:3 offs:1 concurrently:1 amacrine:1 rather:1 voltage:14 properly:1 salamander:1 integrated:1 vlsi:3 relation:1 pixel:1 rodr:5 xps:3 socalled:1 summed:1 fairly:1 field:3 once:1 extraction:1 eliminated:1 mm2:5 identical:1 represents:1 placing:1 mimic:1 stimulus:1 retina:7 oriented:1 neighbour:1 composed:1 phase:1 consisting:3 intended:1 amplifier:1 detection:3 highly:1 possibility:1 intra:2 accurate:1 edge:3 capable:1 vw0:2 re:1 instance:2 unn:1 tp:1 zn:1 deviation:1 reported:1 stored:2 connect:1 supposes:1 density:2 peak:3 international:4 corrects:1 continuously:1 na:2 connectivity:1 pretended:1 corner:1 conf:1 li:2 de:2 retinal:1 includes:1 coefficient:1 int:1 closed:1 wave:6 aggregation:1 parallel:5 contribution:5 accuracy:5 characteristic:1 reina:1 processor:12 synapsis:3 synaptic:5 ed:1 sixth:1 frequency:2 involved:1 pp:7 gain:1 adjusting:1 cicyt:1 lim:1 higher:2 dt:1 permitted:1 synapse:5 done:1 though:1 ano:1 carmona:4 stage:2 traveling:2 receives:1 horizontal:1 nonlinear:2 marker:1 propagation:1 fsr:2 building:1 effect:3 validity:1 multiplier:2 counterpart:1 evolution:5 hence:1 bnn:1 during:2 self:1 spanish:1 m:1 interface:2 image:8 common:1 analog:15 measurement:1 silicon:1 photodetectors:1 ai:1 tuning:1 focal:1 zeroing:1 portugal:1 cancellation:3 csic:1 calibration:1 etc:2 add:1 italy:1 driven:1 forcing:1 periphery:1 store:1 n00014:1 manifested:1 binary:2 onr:1 captured:1 seen:1 additional:2 transmitted:1 employed:1 signal:12 arithmetic:3 full:3 long:1 controlled:1 schematic:1 pitch:1 scalable:1 basic:4 vision:5 represent:1 achieved:2 cell:25 proposal:1 diagram:3 lsb:1 source:1 biased:2 rest:1 operate:3 exhibited:1 markedly:1 induced:1 flow:1 dxn:1 capacitor:4 mw:3 vw:6 presence:1 intermediate:2 feedforward:1 rendering:1 switch:3 architecture:2 converter:1 identified:2 prototype:3 attenuated:1 motivated:1 handled:2 passed:1 cause:1 constitute:1 action:1 programmable:9 governs:1 processed:1 reduced:1 timeconstant:1 notice:1 s3:6 per:8 vol:4 affected:1 ist:1 four:4 mercedes:1 diffusion:2 kept:1 enez:1 year:1 sum:4 run:1 injected:1 communicate:1 solitary:1 place:1 vx0:2 scaling:9 bit:3 layer:24 nan:2 distinguish:1 followed:1 display:1 activity:2 strength:1 multiplexing:1 aspect:1 speed:1 watt:1 across:1 remain:1 smaller:1 ophthalmology:1 son:1 castro:4 invariant:1 sevilla:2 equation:2 bus:2 mechanism:1 needed:1 end:1 operation:5 permit:1 appropriate:1 cancelled:1 neighbourhood:1 alternative:3 robustness:1 gate:2 slower:2 original:1 top:1 remaining:1 llu:1 especially:1 question:1 realized:2 receptive:1 dependence:1 amongst:1 regulated:2 consumption:2 w0:1 cellular:4 code:1 providing:1 negative:1 implementation:3 design:1 proper:1 adjustable:1 upper:1 observation:1 situation:1 precise:1 jim:1 introduced:1 required:2 kl:1 connection:2 roska:2 able:1 below:1 pattern:4 mismatch:2 challenge:1 program:1 built:1 memory:16 power:10 treated:1 lisbon:1 technology:2 eye:1 picture:4 coupled:4 auto:1 speeding:1 drain:1 multiplication:1 fully:2 loss:1 mixed:2 interesting:1 generation:1 digital:10 storing:1 surrounded:1 tiny:1 unscaled:1 row:3 supported:1 last:2 bias:2 side:1 wide:2 taking:1 fifth:1 boundary:1 feedback:2 xn:11 stand:1 dimension:1 contour:1 sensory:1 author:1 preprocessing:2 logic:4 unreliable:1 global:1 active:3 conceptual:1 llm:1 spatio:2 gotarredona:1 continuous:3 table:3 promising:1 nature:1 complex:4 poly:1 domain:1 linearly:1 arrow:1 whole:1 fig:9 referred:1 espejo:4 imse:1 wiley:1 n:1 rail:1 third:1 down:2 offset:10 incorporating:1 workshop:2 roca:1 mirror:7 depicted:2 photograph:1 ganglion:1 visual:3 expressed:1 conveyor:4 tracking:1 partially:1 nominally:1 corresponds:2 relies:1 ann:1 feasible:2 tiger:1 corrected:1 total:1 e:1 experimental:2 rarely:1 internal:3 arises:1 azquez:5 avoiding:1 multiplexed:1 phenomenon:2 |
1,286 | 2,171 | Reinforcement Learning to Play an Optimal
Nash Equilibrium in Team Markov Games
Xiaofeng Wang
ECE Department
Carnegie Mellon University
Pittsburgh, PA 15213
[email protected]
Tuomas Sandholm
CS Department
Carnegie Mellon University
Pittsburgh, PA 15213
[email protected]
Abstract
Multiagent learning is a key problem in AI. In the presence of multiple Nash equilibria, even agents with non-conflicting interests may not
be able to learn an optimal coordination policy. The problem is exaccerbated if the agents do not know the game and independently receive
noisy payoffs. So, multiagent reinforfcement learning involves two interrelated problems: identifying the game and learning to play. In this paper,
we present optimal adaptive learning, the first algorithm that converges
to an optimal Nash equilibrium with probability 1 in any team Markov
game. We provide a convergence proof, and show that the algorithm?s
parameters are easy to set to meet the convergence conditions.
1 Introduction
Multiagent learning is a key problem in AI. For a decade, computer scientists have worked
on extending reinforcement learning (RL) to multiagent settings [11, 15, 5, 17]. Markov
games (aka. stochastic games) [16] have emerged as the prevalent model of multiagent RL.
An approach called Nash-Q [9, 6, 8] has been proposed for learning the game structure
and the agents? strategies (to a fixed point called Nash equilibrium where no agent can
improve its expected payoff by deviating to a different strategy). Nash-Q converges if a
unique Nash equilibrium exists, but generally there are multiple Nash equilibria. Even
team Markov games (where the agents have common interests) can have multiple Nash
equilibria, only some of which are optimal (that is, maximize sum of the agents? discounted
payoffs). Therefore, learning in this setting is highly nontrivial.
A straightforward solution to this problem is to enforce convention (social law). Boutilier
proposed a tie-breaking scheme where agents choose individual actions in lexicographic
order[1]. However, there are many settings where the designer is unable or unwilling
to impose a convention. In these cases, agents need to learn to coordinate. Claus and
Boutilier introduced fictitious play, an equilibrium selection technique in game theory, to
RL. Their algorithm, joint action learner (JAL) [2], guarantees the convergence to a Nash
equilibrium in a team stage game. However, this equilibrium may not be optimal. The
same problem prevails in other equilibrium-selection approaches in game theory such as
adaptive play [18] and the evolutionary model proposed in [7].
In RL, the agents usually do not know the environmental model (game) up front and receive
noisy payoffs. In this case, even the lexicographic approaches may not work because agents
receive noisy payoffs independently and thus may never perceive a tie. Another significant
problem in previous research is how a nonstationary exploration policy (required by RL) affects the convergence of equilibrium selection approaches?which have been studied under
the assumption that agents either always take the best-response actions or make mistakes at
a constant rate. In RL, learning to play an optimal Nash equilibrium in team Markov games
has been posed as one of the important open problems [9]. While there have been heuristic
approaches to this problm, no existing algorithm has been proposed that is guarenteed to
converge to an optimal Nash equilibrium in this setting.
In this paper, we present optimal adaptive learning (OAL), the first algorithm that converge
to an optimal Nash equilibrium with probability 1 in any team Markov game (Section 3).
We prove its convergence, and show that OAL?s parameters are easy to meet the convergence conditions (Section 4).
2 The setting
2.1 MDPs and reinforcement learning (RL)
In a Markov decision problem, there is one agent
in the environment.
A fully observable
where
is
a
finite
state space;
Markov decision problem (MDP) is a tuple
is the space of actions the agent can take;
is a payoff function ( ' ()+*-,
!"#$%&%
is the expected payoff.for
taking
action
in
state
);
and
is
/0 213
41
a transition
function
(
is
the
probability
of
ending
in
state
,
given
that
ac
tion is taken in state ). An agent?s deterministic
policy (aka. strategy) is a mapping
6
from states to actions. We denote by 5
the action that :Apolicy
5 prescribes in state .
@ CB
B
:ED 5 , where : is the pay5 that maximizes 79:<8 ;=?>
The objective is to find a' ()policy
E*2
H
>
G
off at time F , and
is a discount factor. There exists a deterministic optimal
5JI [12]. The Q-function for this policy, K
I , is defined by the set of equations
policy
/ NM/ PO
L 0 1 VUXWZY)[
1 1
R S\ KLI
> 7QARCS/T
KLI
. At any state , the optimal
W]^UXWZY_[
/
KLI
policy chooses
[10].
Reinforcement learning
can
be
viewed
as a sampling method for
estimating KI when the
/
payoff function and/or
transition
function
are
unknown.
KI
can be approximated
/
by a function K :
calculated from the agent?s experience
up
to
time F . The model
based
approach
uses
samples
to
generate
models
of
and
,
and
then
iteratively
computes
/ V`M
/ PO
/ ab 1 VUXWZY [
1 1
:
:
RcS\ K :edf
> 7
.
K :
Q R S/T
Based on K : , a learning policy assigns probabilities to actions at each state. If the learning
policy has the ?Greedy in the Limit with Infinite Exploration? (GLIE) property, then K :
will converge to K.I (with either a model-based or model-free approach) and the agent will
converge in behavior to an optimal policy [14]. Using GLIE, every state-action pair is
visited infinitely often, and in the limit the action selection is greedy with respect to the
Q-function w.p.1. One common GLIE policy is Boltzmann exploration [14].
2.2 Multiagent RL in team Markov games when the game is unknown
A natural extension of an MDP to multiagent environments is a Markov game (aka.
stochastic game) [16]. In this paper we focus on team Markov games, that are Markov
games where each agent receives the same expected payoff (in the presence of noise, different agent may still receive different payoffs at a particular moment.). In other words,
there are no conflicts between the agents, but learning the game structure and learning to
coordinate are nevertheless highly nontrivial.
g is a tuple
Definition
1 A team Markov
game (aka identical-interest stochastic j
game)
Chi
h
M`lk
;mfonpnpn q is a
, where is a setr
ofsn9agents;
S
is
a
finite
state
space;
tuv
jointw
action
space of n agents;
is the common expected payoff function;
Vxy#z{|' (_E*E,
and
is a transition function.
The objective of the
} agents is to find a deterministic joint policy (aka. joint strategy aka.
k
k ??? k
M#~
/{?
strategy profile) 5
5 ;Jfbnpnpn q0 (where 5
and 5
) so as to maximize the
/
expected sum of their discounted payoffs. The Q-function, K
, is the expected sum
of discounted payoffs given that the agents
play
joint
action
in
state
and follow policy
/ V
5 thereafter. The optimal K -function K.I
is the K -function for (each) optimal policy
5mI . So, KLI captures the game structure. The agents generally do not know KI in advance.
Sometimes, they know neither the payoff structure nor the transition probabilities.
f
~
k
if each individual policy1 is a best
response
A joint policy 5 ;mfonpnpn q is a Nashh equilibrium
/-~
k 2
k
G
G
5 , KLI
5
to the
others.
That
is,
for
all
,
and
any
individual
policy
k 2
k 2
k
/-~
1 2
k
d
d
d
5
KLI
5
5
, where 5
is the joint policy of all agents except
agent
.
(Likewise,
throughout
the
paper,
we
use
to denote all agents but , e.g., by
k
k
d
their joint action, by d their joint action set.) A Nash equilibrium is strict if the
inequality above is strict. An optimal Nash equilibrium 5 I is a Nash equilibrium that gives
the agents the maximal expected sum of discounted payoffs. In team games, each optimal
Nash equilibrium is an optimal joint policy (and there are no other optimal joint policies).
/
/ 1
1
G
A joint
action is optimal in state if
for all
. If we treat
K.I
KLI
/
as the payoff of joint action in state , we obtain a team game
in
matrix
form.
K I
We call such a game a state game for . An optimal joint action in is an optimal Nash
equilibrium of that state game. Thus, the task of optimal coordination in a team Markov
game boils down to having all the agents play an optimal Nash equilibrium in state games.
f f
f
f
f
f
However, a coordination problem arises if there are multiple Nash equilibria. The 3-player
[
[
[
f
10
-20
-20
-20
-20
5
-20
5
-20
-20
-20
5
-20
10
-20
5
-20
-20
-20
5
-20
5
-20
-20
-20
-20
10
Table 1: A three-player coordination game
game in Table 1 has three optimal Nash equilibria and six sub-optimal Nash equilibria. In
this game, no existing equilibrium selection algorithm (e.g.,fictitious play [3]) is guarenteed
to learn to play an optimal Nash equilibrium. Furthermore, if the payoffs are only expectations over each agent?s noisy payoffs and unknown to the agents before playing, even
identification of these sub-optimal Nash equilibria during learning is nontrivial.
3 Optimal adaptive learning (OAL) algorithm
We first consider the case where agents know the game before playing.
This enables the
learning agents to construct a virtual game (VG) for each state of the team Markov
game to eliminate all the strict suboptimal Nash equilibria in that state. Let I
be the/payoff
that
the
agents
receive from/the
VG in state for
a joint
action . We let
VyM
*
9M
WZ] ^`UXW4Y)[
1
M
(
R S\ KLI
LI
if
and LI
otherwise. For example,
the VG~6 for the
game in~6Table
1 gives payoff 1 for each optimal Nash equilibrium
~60* 2* 4*
) )
,
, and
), and payoff 0 to every other joint action. The
(
VG in this example is weakly acyclic.
Definition 2 (Weakly acyclic game [18]) Let be an n-player
game in matrix form. The
G
best-response
graph
of
takes
each
joint
action
as
a
vertex
and connects two
1
1
M
1
vertices and with a directed
edge
if
and
only
if
1)
; 2) there exists
k
k
k1
1 k M9
d . We say the game
exactly one agent such that is a best response to d and d
is weakly acyclic if in its best-response graph, from any initial vertex , there exists a
directed path to some vertex I from which there is no outgoing edge.
To tackle the equilibrium selection problem for weakly acyclic games, Young [18]
pro :
G
posed a learning algorithm called adaptive play (AP), which works as follows. Let
be a joint action played at time F over an n-player game in matrix form. Fix integers
Throughout the paper, every Nash equilibrium that we discuss is also a subgame perfect Nash equilibrium. (This refinement
of Nash equilibrium was first introduced in [13] for different games).
. When , each agent randomly chooses its
agent looks back at the most recent plays
and, each
randomly
(without replacement) selects samples
from . Let
be the number of times that a reduced joint action
(a
joint action without agent ?s individual action) appears in the samples at
. Let
be agent ?s payoff given that joint action has been played. Agent calculates its expected
! ,
payoff w.r.t its individual action
as
and then randomly
an action from a set of best responses: "
# chooses
.
*
and
such that
actions.
Starting
from
mf
M? :ed
:ed
:
:
: mf
C
F
M
O
F
*
- :ed?f
d
k
F
@
k
WZ] ^`UXW4Y [
R S\
@
1k
k
d
O{*
CVkeyM
7
[
k ~6Vk
S\
M
k
d
k
[
ke
d
k:
G
~2?k
VkM
D
Young showed that AP in a weakly acyclic game converges to a strict Nash equilibrium
w.p.1. Thus, AP on the VG for the game in Table 1 leads to an equilibrium with payoff 1
which is actually an optimal Nash equilibrium for the original game. Unfortunately, this
does not extend to all VGs because not all VGs are weakly acyclic: in a VG without any
strict Nash equilibrium, AP may not converge to the strategy profile with payoff 1.
In order to address more general settings, we now modify the notion of weakly acyclic
game and adaptive play to accommodate weak optimal Nash equilibria.
Definition 3 (Weakly acyclic game w.r.t a biased set (WAGB)): Let be a set containing some of the Nash equilibria of a game (and no other joint policies). Game is a
WAGB if, from any initial vertex , there exists a directed path to either a Nash equilibrium
inside or a strict Nash equilibrium.
We can convert any VG to a WAGB by setting the biased set to include all joint policies that give payoff 1 (and no other joint policies). To solve such a game, we introduce
a new learning algorithm for equilibrium selection. It enables each agent to deterministically select a best-response action once any Nash equilibrium in the biased set is attained
(even if there exist several best responses when the Nash equilibrium is not strict). This is
different from AP where players randomize their action selection when there are multiple
best-response actions. We call our approach biased adaptive play (BAP). BAP works as
follows. Let
be the biased set composed of some Nash equilibria of a game in matrix
: be the set of
form. Let
samples drawn at time F , without replacement,
from among
1
G
joint
actions.
If
(1)
there
exists
a
joint
action
such
that for all
the
most
recent
k
k
1
: , d
d
G
and
,
and
(2)
there
exists
at
least
one
joint
action
such
that
k
k
: R
: and
G
G
G
,
then
agent
chooses
its
best-response
action
such
that
and
UXW4Ys~6
Vk
1 M
. That is,
D
:
G
G
F
is contained in the most recent play of
a Nash equilibrium inside . On the other hand, if the two conditions above are not met,
then agent chooses its best-response action in the same way as AP. As we will show, BAP
(even with GLIE exploration) on a WAGB converges w.p.1 to either a Nash equilibrium in
or a strict Nash equilibrium.
So far we tackled learning of coordination in team Markov games where the game structure
is known. Our real interests are in learning when the game is unknown. In multiagent
:
reinforcement learning, K
I /is asymptotically approximated with K : . Let
be
: so as to assure
the virtual
game
w.r.t
.
Our
question
is
how
to
construct
K :
:
LI w.p.1. Our method of achieving this makes use of the notion of -optimality.
$
$
$
$
&%
( $
$
'%
*) ( $
$
$
+
+
+
+
+
Definition
4 Let beUXaW4Y)positive
constant. A joint
action a is -optimal at state s and time
/ `O
[
1
1
R K :
G
t if K :
for
all
. We denote the set of -optimal joint
2
actions at state s and time t as
.
2
The idea is to use a decreasing -bound : to estimate
at state and time F . All the
joint actions belonging to the set are treated as optimal Nash equilibria in the virtual game
: which give agents payoff 1. If : converges to zero at a rate slower than K : converges
&,
+
+
+
&,
+
"
' (_E*E,
:
:
G
to KLI , then
LI w.p.1. We make : proportional to a function
:
:
which decreases slowly and monotonically to zero with , where
is the smallest number of times that any state-action pair has been sampled so far. Now, we are ready to present
the entire
optimal
adaptive learning (OAL) algorithm. As we will present thereafter, we
: carefully using an understanding of the convergence rate of a model-based RL
craft
algorithm that is used to learn the game
structure.
k
"
Optimal adaptive learning algorithm (for agent )
,
:J;
=
1. Initialization
;
; \
Q
;
. For all Q ST
\ . ; \ .
[
S
\
and
:
do
q
Q
[
;{f
,
(
[
Q
;
Q R
2. Learning of coordination policy If
, randomly select
[ an action,
[ otherwise do
;yf
Q
[
[
(a) Update the virtual
game
at state Q :
if S \
Q and
;
;
f
Q
Set
.
(b) According to GLIE exploration, with an exploitation probability do
!
Q
and
[
Q
;X=
[
;=
.
otherwise.
i. Randomly select (without replacement) records from
recent observations of others? joint actions
played at state Q .
[
[
Q
)
(
ii. Calculate expected payoff
of individual action '& over
the virtual game
at current state Q
[
[
[
as follows:
Q
;
# %$
"!
7
:
;
response set at state Q and time : ,-
Q
*
[ [
Q
;/.021435.6
+
R
. Construct the best-
"!
Q
[
R
.
iii. If conditions 1) and 2) of BAP are met, choose a best-response action with respect to the biased set .
Q .
Otherwise, randomly select a best-response action from ,-
Otherwise, randomly select an action to explore.
( ( (
(
(
(
,,
,
is the number of times a joint action has been played in state by time .
Here,
is a positive constant (any value works).
is the number of times that a joint
action
appears in agent ?s samples (at time ) from the most recent joint
3. Off-policy learning of game structure 798
[
Q:
Q R and payoff ; under the joint action . Do
(a) Observe state
transition
[ 4
<
[
q
q
.
f
Q [ 4<
Q= [
i.
.
[
(
d
Q
Q
Q
ii.
.[
;
>
$ '&
[
4<
[
M
Q
AQ R
>
/
} :
Q R
( f?d
$ '& [
4<
?A; @ Q R do
Q= Q%?
[
ED
CB
Q= Q R
R
7
[
"D
! q
I
Q
&
.
Q Q R .
(
fd
Q
i
?
S
T
Q
iv. For all
and
> )$ %&
[ 4<
[
35.6 "!
Q=
Q
R
(b) 7
7
<
<
:
: .f
35G H
(c)
.F
(d) If J < , F
(see Section 4.2 for the construction of , F
)
, F [ .
i.
[
ii. Update4
using (b). I!
7 < Q [K for all[ Q
[
Q=
Q= R .
L 35.6 R
7
7
iii. \
Q
Q=
iii.
d
k
G
d
:
k
d
k
Q R
[
[
Q
R
eQ ?
.
.
F
F
actions taken in state .
4 Proof of convergence of OAL
In this section, we prove that OAL converges to an optimal Nash equilibrium. Throughout,
we make the common RL assumptions: payoffs are bounded, and the number of states
and actions is finite. The proof is organized as follows. In Section 4.1 we show that OAL
agents learn optimal coordination if the game is known. Specifically, we show that BAP
against a WAGB with known game structure converges to a Nash equilibrium under GLIE
exploration. Then in Section 4.2 we show that OAL agents will learn the game structure.
Specifically, any virtual game can be converted to a WAGB which will be learned surely.
Finally, these two tracks merge in Section 4.3 which shows that OAL agents will learn the
game structure and optimal coordination. Due to limited space, we omit most proofs. They
can be found at: www.cs.cmu.edu/? sandholm/oal.ps.
4.1 Learning to coordinate in a known game
In this section, we first model our biased adaptive play (BAP) algorithm with best-response
action selection as a stationary Markov chain. In the second half of this section we then
model BAP with GLIE exploration as a nonstationary Markov chain.
4.1.1 BAP as a stationary Markov chain
B
M
Consider
BAP
with
randomly selected initial plays. We take the initial
F
C f
+
as the initial state of the Markov
chain. The definition of the other states is
1
inductive: A successor of state is any state obtained by deleting the left-most element
M
of
and3appending
a new right-most element. The only exception is that all the states
C0 a
with being either a member of the biased set or a strict Nash equilibrium
are grouped into a unique terminal state
. Any state directing to G
is treated as
directly connected to .
1
Let be
the~6state
transition
matrix of the above Markov chain. Let be a successor of ,
M
-
f
q ( } players) be the new element that was appended to the right
and let
1
(
1
(
R
R
of to get . Let
be the transition probability from to . Now,
k
if and only if for each agent , there exists a sample of size in to which is ?s best
response according to the action-selection rule of BAP. Because agent chooses such a
sample with a probability independent of time F , the Markov chain is stationary. Finally,
due to our clustering of
multiple states into a terminal state
, for any state connected
M
R
7
R
to , we have
.
S
In the above model, once the system reaches the terminal state, each agent?s best response
is M to Crepeat
its
most recent action. This is straightforward if in the actual terminal state
00
(which is one of the states that were clustered to form the terminal
state),
is a strict Nash equilibrium. If is only a weak Nash equilibrium (in this case, G
), BAP
biases each agent to choose its most recent action because conditions (1) and (2) of BAP are
satisfied. Therefore, the terminal state
is an absorbing state of the finite Markov chain.
On the other hand, the above analysis shows that
essentially is composed of multiple
absorbing states. Therefore, if agents come into , they will be stuck in a particular state
forever instead of cycling around multiple states in .
in
Theorem 1 Let G be a weakly acyclic game w.r.t. a biased set D. Let L(a) be the length
of the shortest directed path in the best-response graph
of G Cfrom
a joint action
a to
either
M
UXW4Y [
O
an absorbing vertex or a vertex in D, and let
. If
, then,
w.p.1, biased adaptive play in G converges to either a strict Nash equilibrium or a Nash
equilibrium in D.
O
Theorem
1 says that the stationary Markov chain for BAP in a WAGB (given
) has a unique stationary distribution in which only the terminal state appears.
4.1.2 BAP with GLIE exploration as a nonstationary Markov chain
Without knowing game structure, the learners need to use exploration to estimate their payoffs. In this section we show that such exploration does not hurt the convergence of BAP.
We show this by first modeling BAP with GLIE exploration as a non-stationary Markov
chain.
(
$
(
(
(
(
(
$
(
(
(
(
With GLIE exploration, at every time step F , each joint action occurs with positive probability. This means that the system transitions from the state it is in to any of the successor
states with positive probability. On the other hand, the agents? action-selection becomes
increasingly greedy over time. In the limit, with probability one, the transition probabilities
converge to those of BAP with no exploration.
Therefore, we can
model the
learning pro~
U
M
:
cess with a sequence of transition matrices : :<8 ;Jf such that
: :
, where is
8
the transition matrix of the stationary Markov chain describing BAP without exploration.
Akin to how we modeled BAP as a stationary Markov chain above, Young modeled adaptive play (AP) as a stationary
Markov chain [18]. There are two differences. First, unlike AP?s, BAP?s action selection is biased. Second, in Young?s model, it
is possible to have several absorbing states while in our model, at most one absorbing state exists (for any team game, our model
has exactly one absorbing state). This is because we cluster all the absorbing states into one. This allows us to prove our main
convergence theorem.
Our objective here is to show that on a WAGB, BAP with GLIE exploration will converge
to the (?clustered?) terminal state. For that, we use the following lemma (which is a combination of Theorems V4.4 and V4.5 from [4]).
matrix of a stationary Markov chain with a unique
Lemma 2 Let be the finite transition
~
: :<;mf be a sequence of finite transition matrices. Let
stationary distribution . Let
8
M
U
M
: :
:
. If
beU a probability
vector and denote
, then
M
(
8
:
for all .
8
Using this lemma and Theorem 1, we can prove the following theorem.
Theorem
3 O (BAP
with GLIE) On a WAGB G, w.p.1, BAP with GLIE exploration (and
/
) converges to either a strict Nash equilibrium or a Nash equilibrium in D.
4.2 Learning the virtual game
So far, we have shown that if the game structure is known in a WAGB, then BAP will
converge to the terminal state. To prove optimal convergence of the OAL algorithm, we
need to further demonstrate that 1) every virtual game is a WAGB, and 2) in OAL, the
: will converge to the ?correct? virtual game
?temporary? virtual game
I w.p.1.
The first of these two issues is handled by the following lemma:
Lemma 4 The virtual game VG of any n-player team state game is a weakly acyclic game
w.r.t a biased set that contains all the optimal Nash equilibria, and no other joint actions.
(By the definition of a virtual game, there are no strict Nash equilibria other than optimal
ones.) The length of the shortest best-response path
} .
Lemma 4 implies that BAP in a known virtual game with GLIE exploration will converge
to an optimal Nash equilibrium. This is because (by Theorem 3) BAP in a WAGB will
converge to either a Nash equilibrium in a biased set or a strict Nash equilibrium, and
(by Lemma 4) any virtual game is a WAGB with all such Nash equilibria being optimal.
$
The following two lemmas are the last link of our proof chain. They show that OAL will
cause agents to obtain the correct virtual game almost surely.
,
Lemma 5 In any team Markov game, (part 3 of) OAL assures that as F
UXW4Y
Q S/T
[
S\
DK
:
?KLI
/ V
D
'
M
1
F
1
F
for some constant
M
(
w.p.1.
Using Lemma 5, the following lemma is easy to prove.
1
M
: R
the
event that for all F F ,
Lemma 6 Consider any team Markov game. Let : be
:
LI in the OAL algorithm in a given state. If 1)
decreases monotonically to zero
(
U
: :
8
"
`M9(
:
), and 2)
U
: :
8
1
,
"
F
1
F
F
M9(
, then
U
: :
8
B?~
+
:
M#*
.
Lemma 6 states that if the criterion for including a joint action among the -optimal joint
actions in OAL is not made strict too quickly (quicker than the iterated logarithm), then
agents will identify all optimal joint actions with probability one. In this case, they set up
the correct virtual
game. It is easy to make OAL satisfy this condition. E.g., any function
d
f
NM
(
:
:
, will do.
4.3 Main convergence theorem
Now we are ready to prove that OAL converges to an optimal Nash equilibrium in any
team Markov game, even when the game structure is unknown. The idea is to show that
the OAL agents learn the game structure (VGs) and the optimal coordination policy (over
these VGs). OAL tackles these two learning problems simultaneously?specifically, it
interleaves BAP (with GLIE exploration) with learning of game structure. However, the
convergence proof does not make use of this fact. Instead, the proof proceeds by showing
that the VGs are learned first, and coordination second (the learning algorithm does not
even itself know when the switch occurs, but it does occur w.p.1).
"
"
Theorem
7 (Optimal
convergence)
In any team Markov game among } agents, if (1)
O
: satisfies Lemma 6, then the OAL algorithm converges to an
, and (2)
}
optimal Nash equilibrium w.p.1.
Proof. According to [1], a team Markov game can be decomposed into a sequence of state
games. The optimal equilibria of these state games form the optimal policy 5I for the
game. By the definition of GLIE exploration, each state in the finite state space will be
visited infinitely often w.p.1. Thus, it is sufficient to only prove that the OAL algorithm
will converge to the optimal policy over individual state games w.p.1.
M
1
: R
F . Let f be any positive
Let : be the event that
LI at that state for all F
L
f
constant. If B?Condition
(2) of the theorem
is satisfied, by Lemma 6 there exists a time
~
*
.
:b
f if F
f .
such that
If : occurs and Condition (1) of the theorem is satisfied, by Theorem 3, OAL will converge
to either a strict Nash equilibrium or a Nash equilibrium in the biased set w.p.1. Furthermore, by Lemma 4, we know that the biased set contains all of the optimal Nash equilibria
(and nothing else), and there are no strict Nash equilibria outside the biased set. Therefore,
if : occurs, then OAL converges to an optimal Nash equilibrium w.p.1. Let
be any
positive constant, and
let
be
the
event
that
the
agents
play
an
optimal
joint
action
at a
1
given state for all
F
.
With
this
notation,
we
can
reword
the
previous
sentence:
there
B?~
*
exists a time
.
F such that if
F , then
D :
+
+
+
+
+
+
+
+
B?~
Put
there exists
a time
+ f +
such that if
+ f +
, then
B?~ together,
B?~
*
- *
*
:
D :
+ f +
+ f +
. Because + f and +
are only used
in the proof (they are not parameters of the OAL algorithm), we can choose them to be
arbitrarily small. Therefore, OAL converges to an optimal Nash equilibrium w.p.1.
5 Conclusions and future research
With multiple Nash equilibria, multiagent RL becomes difficult even when agents do not
have conflicting interests. In this paper, we present OAL, the first algorithm that converges
to an optimal Nash equilibrium with probability 1 in any team Markov game. In the future
work, we consider extending the algorithm to some general-sum Markov games.
Acknowledgments
Wang is supported by NSF grant IIS-0118767, the DARPA OASIS program, and the PASIS
project at CMU. Sandholm is supported by NSF CAREER Award IRI-9703122, and NSF
grants IIS-9800994, ITR IIS-0081246, and ITR IIS-0121678.
References
[1]
[2]
[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
C.Boutilier. Planning, learning and coordination in multi-agent decision processes. In TARK, 1996.
C.Claus and C.Boutilier. The dynamics of reinforcement learning in cooperative multi-agent systems. In AAAI, 1998.
D.Fudenberg and D.K.Levine. The theory of learning in games. MIT Press, 1998.
D.L.Isaacson and R.W.Madsen. Markov chain: theory and applications. John Wiley and Sons, Inc, 1976.
G.Wei . Learning to coordinate actions in multi-agent systems. In IJCAI, 1993.
J.Hu and W.P.Wellman. Multiagent reinforcement learning: theoretical framework and an algorithm. In ICML, 1998.
M.Kandori, G.J.Mailath, and R.Rob. Learning, mutation, and long run equilibria in games. Econometrica, 61(1):29?56,
1993.
M.Littman. Friend-or-Foe Q-learning in general sum game. In ICML, 2001.
M.L.Littman. Value-function reinforcement learning in markov games. J. of Cognitive System Research, 2:55?66, 2000.
M.L.Purterman. Markov decision processes-discrete stochastic dynamic programming. John Wiley, 1994.
M.Tan. Multi-agent reinforcement learning: independent vs. cooperative agents. In ICML, 1993.
R.A.Howard. Dynamic programming and Markov processes. MIT Press, 1960.
R. Selten. Spieltheoretische behandlung eines oligopolmodells mit nachfragetr?agheit. Zeitschrift f?ur die gesamte Staatswissenschaft, 12:301?324, 1965.
S. Singh, T.Jaakkola, M.L.Littman, and C.Szepesvari. Convergence results for single-step on-policy reinforcement learning
algorithms. Machine Learning, 2000.
S.Sen, M.Sekaran, and J. Hale. Learning to coordinate without sharing information. In AAAI, 1994.
F. Thusijsman. Optimality and equilibrium in stochastic games. Centrum voor Wiskunde en Informatica, 1992.
T.Sandholm and R.Crites. Learning in the iterated prisoner?s dilemma. Biosystems, 37:147?166, 1995.
H. Young. The evolution of conventions. Econometrica, 61(1):57?84, 1993.
[14]
[15]
[16]
[17]
[18]
! !
Theorem 3 requires
L
Lemma 4, we do have
L
.
. If Condition (1) of our main theorem is satisfied (
! '
L
q
), then by
| 2171 |@word exploitation:1 open:1 hu:1 accommodate:1 moment:1 initial:5 contains:2 existing:2 current:1 john:2 enables:2 update:1 v:1 stationary:11 greedy:3 half:1 selected:1 record:1 prove:8 inside:2 introduce:1 expected:9 behavior:1 nor:1 planning:1 multi:4 chi:1 terminal:9 discounted:4 decreasing:1 decomposed:1 actual:1 becomes:2 project:1 estimating:1 bounded:1 notation:1 maximizes:1 guarantee:1 every:5 tackle:2 tie:2 exactly:2 grant:2 omit:1 before:2 positive:6 scientist:1 treat:1 modify:1 limit:3 mistake:1 zeitschrift:1 meet:2 path:4 ap:8 merge:1 initialization:1 studied:1 limited:1 directed:4 unique:4 acknowledgment:1 subgame:1 word:1 get:1 selection:12 put:1 www:1 deterministic:3 straightforward:2 iri:1 starting:1 independently:2 ke:1 unwilling:1 identifying:1 assigns:1 perceive:1 rule:1 notion:2 coordinate:5 hurt:1 construction:1 play:19 tan:1 programming:2 us:1 pa:2 assure:1 element:3 approximated:2 centrum:1 cooperative:2 levine:1 quicker:1 wang:2 capture:1 calculate:1 connected:2 voor:1 decrease:2 environment:2 nash:68 econometrica:2 littman:3 dynamic:3 prescribes:1 weakly:10 singh:1 dilemma:1 learner:2 po:2 joint:42 darpa:1 outside:1 emerged:1 posed:2 heuristic:1 solve:1 say:2 otherwise:5 noisy:4 itself:1 sequence:3 sen:1 maximal:1 convergence:15 cluster:1 p:1 extending:2 ijcai:1 perfect:1 converges:15 andrew:1 ac:1 friend:1 eq:1 c:3 involves:1 come:1 implies:1 convention:3 met:2 correct:3 stochastic:5 exploration:19 successor:3 virtual:16 fix:1 clustered:2 extension:1 around:1 equilibrium:78 mapping:1 cb:2 smallest:1 visited:2 coordination:11 grouped:1 mit:3 lexicographic:2 always:1 jaakkola:1 focus:1 vk:3 selten:1 prevalent:1 aka:6 eliminate:1 entire:1 selects:1 issue:1 among:3 construct:3 never:1 having:1 once:2 sampling:1 identical:1 look:1 icml:3 future:2 others:2 randomly:8 composed:2 simultaneously:1 individual:7 deviating:1 connects:1 replacement:3 interest:5 policy1:1 highly:2 wellman:1 beu:1 chain:16 tuple:2 edge:2 experience:1 iv:1 logarithm:1 theoretical:1 biosystems:1 modeling:1 vertex:7 front:1 too:1 chooses:6 v4:2 off:2 vgs:5 ym:2 quickly:1 together:1 aaai:2 nm:2 satisfied:4 containing:1 choose:4 prevails:1 slowly:1 cognitive:1 li:6 m9:3 converted:1 inc:1 satisfy:1 tion:1 mutation:1 appended:1 likewise:1 identify:1 weak:2 identification:1 iterated:2 foe:1 reach:1 sharing:1 ed:6 definition:7 against:1 proof:9 mi:1 boil:1 sampled:1 organized:1 carefully:1 actually:1 back:1 appears:3 attained:1 follow:1 response:19 wei:1 furthermore:2 stage:1 hand:3 receives:1 yf:1 mdp:2 inductive:1 evolution:1 q0:1 iteratively:1 game:103 during:1 die:1 criterion:1 isaacson:1 demonstrate:1 pro:2 glie:16 common:4 absorbing:7 rl:11 ji:1 extend:1 mellon:2 significant:1 ai:2 aq:1 interleaf:1 recent:7 showed:1 madsen:1 inequality:1 arbitrarily:1 impose:1 surely:2 converge:13 shortest:2 maximize:2 monotonically:2 ii:7 multiple:9 long:1 rcs:1 award:1 calculates:1 essentially:1 cmu:4 expectation:1 sometimes:1 receive:5 else:1 biased:16 unlike:1 strict:17 claus:2 member:1 call:2 nonstationary:3 integer:1 presence:2 iii:3 easy:4 switch:1 affect:1 suboptimal:1 idea:2 knowing:1 itr:2 six:1 handled:1 akin:1 wiskunde:1 jal:1 cause:1 action:65 boutilier:4 generally:2 discount:1 informatica:1 reduced:1 generate:1 exist:1 nsf:3 designer:1 track:1 carnegie:2 discrete:1 thereafter:2 key:2 nevertheless:1 achieving:1 drawn:1 oal:27 neither:1 ce:1 graph:3 asymptotically:1 sum:6 convert:1 run:1 throughout:3 almost:1 decision:4 bound:1 played:4 tackled:1 nontrivial:3 occur:1 worked:1 optimality:2 department:2 according:3 combination:1 belonging:1 sandholm:5 bap:27 increasingly:1 son:1 ur:1 rob:1 tark:1 taken:2 equation:1 assures:1 discus:1 describing:1 know:7 observe:1 enforce:1 appending:1 slower:1 original:1 clustering:1 include:1 k1:1 objective:3 question:1 occurs:4 strategy:6 randomize:1 cycling:1 evolutionary:1 unable:1 link:1 tuomas:1 length:2 modeled:2 difficult:1 unfortunately:1 policy:28 unknown:5 boltzmann:1 observation:1 markov:41 howard:1 finite:7 payoff:31 team:22 directing:1 introduced:2 pair:2 required:1 sentence:1 conflict:1 learned:2 conflicting:2 temporary:1 address:1 able:1 proceeds:1 usually:1 program:1 wz:2 including:1 deleting:1 event:3 natural:1 treated:2 scheme:1 improve:1 mdps:1 lk:1 ready:2 sn:1 understanding:1 law:1 multiagent:10 fully:1 proportional:1 fictitious:2 acyclic:10 vg:8 agent:66 sufficient:1 playing:2 supported:2 last:1 free:1 bias:1 taking:1 calculated:1 transition:13 ending:1 computes:1 stuck:1 made:1 reinforcement:10 adaptive:12 refinement:1 far:3 social:1 observable:1 forever:1 pittsburgh:2 decade:1 table:4 learn:8 szepesvari:1 career:1 main:3 fudenberg:1 crites:1 noise:1 profile:2 nothing:1 en:1 wiley:2 sub:2 deterministically:1 breaking:1 young:5 down:1 theorem:15 xiaofeng:2 showing:1 hale:1 dk:1 exists:12 mf:3 interrelated:1 explore:1 infinitely:2 contained:1 prisoner:1 oasis:1 environmental:1 satisfies:1 viewed:1 kli:10 jf:1 infinite:1 except:1 specifically:3 lemma:17 called:3 ece:1 player:7 craft:1 exception:1 select:5 arises:1 outgoing:1 |
1,287 | 2,172 | VIBES: A Variational Inference
Engine for Bayesian Networks
Christopher M. Bishop
Microsoft Research
Cambridge, CB3 0FB, U.K.
research.microsoft.com/?cmbishop
David Spiegelhalter
MRC Biostatistics Unit
Cambridge, U.K.
[email protected]
John Winn
Department of Physics
University of Cambridge, U.K.
www.inference.phy.cam.ac.uk/jmw39
Abstract
In recent years variational methods have become a popular tool
for approximate inference and learning in a wide variety of probabilistic models. For each new application, however, it is currently
necessary first to derive the variational update equations, and then
to implement them in application-specific code. Each of these steps
is both time consuming and error prone. In this paper we describe a
general purpose inference engine called VIBES (?Variational Inference for Bayesian Networks?) which allows a wide variety of probabilistic models to be implemented and solved variationally without
recourse to coding. New models are specified either through a
simple script or via a graphical interface analogous to a drawing
package. VIBES then automatically generates and solves the variational equations. We illustrate the power and flexibility of VIBES
using examples from Bayesian mixture modelling.
1
Introduction
Variational methods [1, 2] have been used successfully for a wide range of models,
and new applications are constantly being explored. In many ways the variational
framework can be seen as a complementary approach to that of Markov chain Monte
Carlo (MCMC), with different strengths and weaknesses.
For many years there has existed a powerful tool for tackling new problems using
MCMC, called BUGS (?Bayesian inference Using Gibbs Sampling?) [3]. In BUGS
a new probabilistic model, expressed as a directed acyclic graph, can be encoded
using a simple scripting notation, and then samples can be drawn from the posterior
distribution (given some data set of observed values) using Gibbs sampling in a way
that is largely automatic. Furthermore, an extension called WinBUGS provides a
graphical front end to BUGS in which the user draws a pictorial representation of
the directed graph, and this automatically generates the required script.
We have been inspired by the success of BUGS to produce an analogous tool for
the solution of problems using variational methods. The challenge is to build a
system that can handle a wide range of graph structures, a broad variety of common conditional probability distributions at the nodes, and a range of variational
approximating distributions. All of this must be achieved whilst also remaining
computationally efficient.
2
A General Framework for Variational Inference
In this section we briefly review the variational framework, and then we characterise a large class of models for which the variational method can be implemented
automatically. We denote the set of all variables in the model by W = (V, X) where
V are the visible (observed) variables and X are the hidden (latent) variables. As
with BUGS, we focus on models that are specified in terms of an acyclic directed
graph (treatment of undirected graphical models is equally possible and is somewhat more straightforward). The joint distribution P (V, X) is then expressed in
terms of conditional distributions P (Wi |pai ) at each node i, where pai denotes the
set of variables corresponding to the parents of node i, and Wi denotes the variable,
or group of variables, associated with node i. The joint distribution
of all variables
Q
is then given by the product of the conditionals P (V, X) = i P (Wi |pai ).
Our goal is to find a variational distribution Q(X|V ) that approximates the true
posterior distribution P (X|V ). To do this we note the following decomposition of
the log marginal probability of the observed data, which holds for any choice of
distribution Q(X|V )
ln P (V ) = L(Q) + KL(QkP )
(1)
where
L(Q)
=
X
Q(X|V ) ln
X
KL(QkP )
= ?
X
X
P (V, X)
Q(X|V )
Q(X|V ) ln
P (X|V )
Q(X|V )
(2)
(3)
and the sums are replaced by integrals in the case of continuous variables. Here
KL(QkP ) is the Kullback-Leibler divergence between the variational approximation
Q(X|V ) and the true posterior P (X|V ). Since this satisfies KL(QkP ) ? 0 it follows
from (1) that the quantity L(Q) forms a lower bound on ln P (V ).
We now choose some family of distributions to represent Q(X|V ) and then seek a
member of that family that maximizes the lower bound L(Q). If we allow Q(X|V )
to have complete flexibility then we see that the maximum of the lower bound
occurs for Q(X|V ) = P (X|V ) so that the variational posterior distribution equals
the true posterior. In this case the Kullback-Leibler divergence vanishes and L(Q) =
ln P (V ). However, working with the true posterior distribution is computationally
intractable (otherwise we wouldn?t be resorting to variational methods). We must
therefore consider a more restricted family of Q distributions which has the property
that the lower bound (2) can be evaluated and optimized efficiently and yet which
is still sufficiently flexible as to give a good approximation to the true posterior
distribution.
2.1
Factorized Distributions
For the purposes of building VIBES we have focussed attention initially on distributions that factorize with respect to disjoint groups Xi of variables
Y
Q(X|V ) =
Qi (Xi ).
(4)
i
This approximation has been successfully used in many applications of variational
methods [4, 5, 6]. Substituting (4) into (2) we can maximize L(Q) variationally
with respect to Qi (Xi ) keeping all Qj for j 6= i fixed. This leads to the solution
ln Q?i (Xi ) = hln P (V, X)i{j6=i} + const.
(5)
where h?ik denotes an expectation with respect to the distribution Qk (Xk ). Taking
exponentials of both sides and normalizing we obtain
Q?i (Xi ) = P
exphln P (V, X)i{j6=i}
.
Xi exphln P (V, X)i{j6=i}
(6)
Note that these are coupled equations since the solution for each Qi (Xi ) depends on
expectations with respect to the other factors Qj6=i . The variational optimization
proceeds by initializing each of the Qi (Xi ) and then cycling through each factor in
turn replacing the current distribution with a revised estimate given by (6). The
current version of VIBES is based on a factorization of the form (4) in which each
factor Qi (Xi ) corresponds to one of the nodes of the graph (each of which can be a
composite node, as discussed shortly).
An important property of the variational update equations, from the point of view of
VIBES, is that the right hand side of (6) does not depend on all of the conditional
distributions P (Wi |pai ) that define the joint distribution but only on those that
have a functional dependence on Xi , namely the conditional P (Xi |pai ), together
with the conditional distributions for any children of node i since these have X i in
their parent set. Thus the expectations that must be performed on the right hand
side of (6) involve only those variables lying in the Markov blanket of node i, in
other words the parents, children and co-parents of i, as illustrated in Figure 1(a).
This is a key concept in VIBES since it allows the variational update equations to
be expressed in terms of local operations, which can therefore be expressed in terms
of generic code which is independent of the global structure of the graph.
2.2
Conjugate Exponential Models
It has already been noted [4, 5] that important simplifications to the variational
update equations occur when the distributions of the latent variables, conditioned
on their parameters, are drawn from the exponential family and are conjugate with
respect to the prior distributions of the parameters. Here we adopt a somewhat different viewpoint in that we make no distinction between latent variables and model
parameters. In a Bayesian setting these both correspond to unobserved stochastic
variables and can be treated on an equal footing. This allows us to consider conjugacy not just between variables and their parameters, but hierarchically between
all parent-child pairs in the graph.
Thus we consider models in which each conditional distribution takes the standard
exponential family form
ln P (Xi |Y ) = ?i (Y )T ui (Xi ) + fi (Xi ) + gi (Y )
(7)
where the vector ?(Y ) is called the natural parameter of the distribution. Now
(i)
consider a node Zj with parent Xi and co-parents cpj , as indicated in Figure 1(a).
Y1
YK
cpj(i)
{
Xi
Z1
}
Zj
(a)
(b)
Figure 1: (a) A central observation is that the variational update equations for node
Xi depend only on expectations over variables appearing in the Markov blanket of
Xi , namely the set of parents, children and co-parents. (b) Hinton diagram of hW i
from one of the components in the Bayesian PCA model, illustrating how all but
three of the PCA eigenvectors have been suppressed.
As far as the pair of nodes Xi and Zj are concerned, we can think of P (Xi |Y )
(i)
as a prior over Xi and the conditional P (Zj |Xi , cpj ) as a (contribution to) the
likelihood function. Conjugacy requires that, as a function of Xi , the product
of these two conditionals must take the same form as (7). Since the conditional
(i)
P (Zj |Xi , cpj ) is also in the exponential family it can be expressed as
(i)
(i)
(i)
ln P (Zj |Xi , cpj ) = ?j (Xi , cpj )T uj (Zj ) + fj (Zj ) + gj (Xi , cpj ).
(8)
Conjugacy then requires that this be expressible in the form
(i)
(i)
(i)
ln P (Zj |Xi , cpj ) = ?ej?i (Zj , cpj ) T ui (Xi ) + ?(Zj , cpj )
(9)
for some choice of functions ?e and ?. Since this must hold for each of the parents of
(i)
Zj it follows that ln P (Zj |Xi , cpj ) must be a multi-linear function of the uk (Xk ) for
each of the parents Xk of node XZj . Also, we observe from (8) that the dependence
(i)
of ln P (Zj |Xi , cpj ) on Zj is again linear in the function uj (Zj ). We can apply a
similar argument to the conjugate relationship between node Xj and each of its
parents, showing that the contribution from the conditional P (Xi |Y ) can again be
expressed in terms of expectations of the natural parameters for the parent node
distributions. Hence the right hand side of the variational update equation (5) for
a particular node Xi will be a multi-linear function of the expectations hui for each
node in the Markov blanket of Xi .
The variational update equation then takes the form
?
?T
M
?
?
X
(i)
ln Q?i (Xi ) = h?i (Y )iY +
h?ej?i (Zj , cpj )iZj ,cp(i)
ui (Xi ) + const.
j ?
?
(10)
j=1
which involves summation of bottom up ?messages? h?ej?i iZj ,cp(i) from the children
j
together with a top-down message h?i (Y )iY from the parents. Since all of these
messages are expressed in terms of the same basis ui (Xi ), we can write compact,
generic code for updating any type of node, instead of having to take account
explicitly of the many possible combinations of node types in each Markov blanket.
As an example, consider the Gaussian N (X|?, ? ?1 ) for a single variable X with
mean ? and precision (inverse variance) ? . The natural coordinates are uX =
[X, X 2 ]T and the natural parameterization is ? = [??, ?? /2]T . Then hui = [?, ?2 +
? ?1 ]T , and the function fi (Xi ) is simply zero in this case. Conjugacy allows us to
choose a distribution for the parent ? that is Gaussian and a prior for ? that is
a Gamma distribution. The corresponding natural parameterizations and update
messages are given by
?
h? ihXi
?
?h(X ? ?)2 i
e
e
,
h
?
i
=
,
u
=
,
h
?
i
=
.
u? =
X??
?
X??
?h? i/2
ln ?
1/2
?2
We can similarly consider multi-dimensional Gaussian distributions, with a Gaussian prior for the mean and a Wishart prior for the inverse covariance matrix.
A generalization of the Gaussian is the rectified Gaussian which is defined as
P (X|?, ? ) ? N (X|?, ? ) for X ? 0 and P (X|?, ? ) = 0 for X < 0, for which moments
can be expressed in terms of the ?erf? function. This rectification corresponds to
the introduction of a step function, whose logarithm corresponds to fi (Xi ) in (7),
which is carried through the variational update equations unchanged. Similarly, we
can consider doubly truncated Gaussians, which are non-zero only over some finite
interval.
Another example is the discrete distribution for categorical variables. These are
most conveniently represented using the 1-of-K scheme in which S = {Sk } with
P
QK
k = 1, . . . , K, Sk ? {0, 1} and k Sk = 1. This has distribution P (S|?) = k=1 ?kSk
and we can place a conjugate Dirichlet distribution over the parameters {?k }.
2.3
Allowable Distributions
We now characterize the class of models that can be solved by VIBES using the
factorized variational distribution given by (4). First of all we note that, since a
Gaussian variable can have a Gaussian parent for its mean, we can extend this hierarchically to any number of levels to give a sub-graph which is a DAG of Gaussian
nodes of arbitrary topology. Each Gaussian can have Gamma (or Wishart) prior
over its precision.
Next, we observe that discrete variables S = {Sk } can be used to construct ?pick?
functions which choose a particular parent node Yb from amongst several conjugate
QK
parents {Yk }, so that Yb = Yk when sk = 1, which can be written Yb = k=1 YkSk .
QK
Under any non-linear function h(?) we have h(Y )P= k=1 h(Yk )Sk . Furthermore the
expectation under S takes the form hh(Y )iS = k hSk ih(Yk ). Variational inference
will therefore be tractable for this model provided it is tractable for each of the
parents Yk individually.
Thus we can handle the following very general architecture: an arbitrary DAG
of multinomial discrete variables (each having Dirichlet priors) together with an
arbitrary DAG of linear Gaussian nodes (each having Wishart priors) and with
arbitrary pick links from the discrete nodes to the Gaussian nodes. This graph
represents a generalization of the Gaussian mixture model, and includes as special
cases models such as hidden Markov models, Kalman filters, factor analysers and
principal component analysers, as well as mixtures and hierarchical mixtures of all
of these.
There are other classes of models that are tractable under this scheme, for example
Poisson variables having Gamma priors, although these may be of limited interest.
We can further extend the class of tractable models by considering nodes whose
natural parameters are formed from deterministic functions of the states of several
parents. This is a key property of the VIBES approach which, as with BUGS, greatly
extends its applicability. Suppose we have some conditional distribution P (X|Y, . . .)
and we want to make Y some deterministic function of the states of some other nodes
?(Z1 , . . . , ZM ). In effect we have a pseudo-parent that is a deterministic function of
other nodes, and indeed is represented explicitly through additional deterministic
nodes in the graphical interface both to WinBUGS and to VIBES. This will be
tractable under VIBES provided the expectation of u? (?) can be expressed in terms
of the expectations of the corresponding functions uj (Zj ) of the parents. The pick
functions discussed earlier are a special case of these deterministic functions.
Thus for a Gaussian node the mean can be formed from products and sums of the
states of other Gaussian nodes provided the function is linear with respect to each
of the nodes. Similarly, the precision of the Gaussian can comprise the products
(but not sums) of any number of Gamma distributed variables.
Finally, we have seen that continuous nodes can have both discrete and continuous
parents but that discrete nodes can only have discrete parents. We can allow discrete
nodes to have continuous parents by stepping outside the conjugate-exponential
framework by exploiting a variational bound on the logistic sigmoid function [1].
We also wish to be able to evaluate the lower bound (2), both to confirm the
correctness of the variational updates (since the value of the bound should never
decrease), as well as to monitor convergence and set termination criteria. This can
be done efficiently, largely using quantities that have already been calculated during
the variational updates.
3
VIBES: A Software Implementation
Creation of a model in VIBES simply involves drawing the graph (using operations
similar to those in a simple drawing package) and then assigning properties to each
node such as the functional form for the distribution, a list of the other variables
it is conditioned on, and the location of the corresponding data file if the node is
observed. The menu of distributions available to the user is dynamically adjusted
at each stage to ensure that only valid conjugate models can be constructed.
As in WinBUGS we have adopted the convention of making logical (deterministic)
nodes explicit in the graphical representation as this greatly simplifies the specification and interpretation of the model. We also use the ?plate? notation of a
box surrounding one or more nodes to denote that those nodes are replicated some
number of times as specified by the parameter appearing in the bottom right hand
corner of the box.
3.1
Example: Bayesian Mixture Models
We illustrate VIBES using a Bayesian model for a mixture of M probabilistic PCA
distributions, each having maximum intrinsic dimensionality of q, with a sparse
prior [6], for which the VIBES implementation is shown in Figure 2. Here there are
N observations of the vector t whose dimensionality is d, as indicated by the plates.
The dimensionality of the other variables is also determined by which plates they
are contained in (e.g. W has dimension d ? q ? M whereas ? is a scalar). Variables
t, x, W and ? are Gaussian, ? and ? have Gamma distributions, S is discrete and
? is Dirichlet.
Once the model is completed (and the file or files containing the observed variables
Figure 2: Screen shot from VIBES showing the graph for a mixture of probabilistic
PCA distributions. The node t is coloured black to denote that this variable is
observed, and the node ?alpha? has been highlighted and its properties (e.g. the
form of the distribution) can be changed using the menus on the left hand side.
The node labelled ?x.W+mu? is a deterministic node, and the double arrows denote
deterministic relationships.
are specified) it is then ?compiled?, which involves allocation of memory for the
variables and initializing the distributions Qi (which is done using simple heuristics
but which can also be over-ridden by the user). If desired, monitoring of the lower
bound (2) can be switched on (at the expense of slightly increased computation) and
this can also be used to set a termination criterion. Alternatively the variational
optimization can be run for a fixed number of iterations.
Once the optimization is complete various diagnostics can be used to probe the
results, such as the Hinton diagram plot shown in Figure 1(b).
Now suppose we wish to modify the model, for instance by having a single set of
hyper-parameters ? whose values are shared by all of the M components in the
mixture, instead of having a separate set for each component. This simply involved
dragging the ? node outside of the M plate using the mouse and then recompiling
(since ? is now a vector of length q instead of a matrix of size M ? q). This literally
takes a few seconds, in contrast to the effort required to formulate the variational
inference equations, and develop bespoke code, for a new model! The result is then
optimized as before. A screen shot of the corresponding VIBES model is shown in
Figure 3.
4
Discussion
Our early experiences with VIBES have shown that it dramatically simplifies the
construction and testing of new variational models, and readily allows a range of
alternative models to be evaluated on a given problem. Currently we are extending
VIBES to cater for a broader range of variational distributions by allowing the user
to specify a Q distribution defined over a subgraph of the true graph [7].
Finally, there are many possible extensions to the basic VIBES we have described
Figure 3: As in Figure 2 but with the vector ? of hyper-parameters moved outside
the M ?plate?. This causes there to be only q terms in ? which are shared over the
mixture components rather than M ? q. Note that, with no nodes highlighted, the
side menus disappear.
here. For example, in order to broaden the range of models that can be tackled we
can combine variational with other methods such as Gibbs sampling or optimization
(empirical Bayes) to allow for non-conjugate hyper-priors for instance. Similarly,
there is scope for exploiting exact methods where there exist tractable sub-graphs.
References
[1] M. I. Jordan, Z. Ghahramani, T. S. Jaakkola, and L. K. Saul. An introduction
to variational methods for graphical models. In M. I. Jordan, editor, Learning
in Graphical Models, pages 105?162. Kluwer, 1998.
[2] R. M. Neal and G. E. Hinton. A new view of the EM algorithm that justifies
incremental and other variants. In M. I. Jordan, editor, Learning in Graphical
Models, pages 355?368. Kluwer, 1998.
[3] D J Lunn, A Thomas, N G Best, and D J Spiegelhalter. WinBUGS ? a Bayesian
modelling framework: concepts, structure and extensibility. Statistics and Computing, 10:321?333, 2000. http://www.mrc-bsu.cam.ac.uk/bugs/.
[4] Z. Ghahramani and M. J. Beal. Propagation algorithms for variational Bayesian
learning. In T. K. Leen, T. Dietterich, and V. Tresp, editors, Advances in Neural
Information Processing Systems, volume 13, Cambridge MA, 2001. MIT Press.
[5] H. Attias. A variational Bayesian framework for graphical models. In S. Solla,
T. K. Leen, and K-L Muller, editors, Advances in Neural Information Processing
Systems, volume 12, pages 209?215, Cambridge MA, 2000. MIT Press.
[6] C. M. Bishop. Variational principal components. In Proceedings Ninth International Conference on Artificial Neural Networks, ICANN?99, volume 1, pages
509?514. IEE, 1999.
[7] Christopher M. Bishop and John Winn. Structured variational distributions in
VIBES. In Proceedings Artificial Intelligence and Statistics, Key West, Florida,
2003. Accepted for publication.
| 2172 |@word illustrating:1 version:1 briefly:1 termination:2 seek:1 decomposition:1 covariance:1 pick:3 shot:2 moment:1 phy:1 ridden:1 current:2 com:1 tackling:1 yet:1 must:6 written:1 assigning:1 john:2 readily:1 visible:1 plot:1 update:11 intelligence:1 parameterization:1 xk:3 footing:1 provides:1 parameterizations:1 node:44 location:1 constructed:1 become:1 ik:1 doubly:1 combine:1 indeed:1 multi:3 inspired:1 automatically:3 considering:1 provided:3 notation:2 maximizes:1 biostatistics:1 factorized:2 whilst:1 unobserved:1 cmbishop:1 pseudo:1 uk:4 unit:1 before:1 local:1 modify:1 black:1 dynamically:1 co:3 factorization:1 limited:1 range:6 directed:3 testing:1 implement:1 empirical:1 composite:1 word:1 www:2 deterministic:8 straightforward:1 attention:1 formulate:1 menu:3 handle:2 coordinate:1 analogous:2 qkp:4 construction:1 suppose:2 user:4 exact:1 updating:1 observed:6 bottom:2 solved:2 initializing:2 solla:1 decrease:1 extensibility:1 yk:6 vanishes:1 mu:1 ui:4 cam:3 depend:2 creation:1 basis:1 joint:3 represented:2 various:1 surrounding:1 describe:1 monte:1 artificial:2 analyser:2 hyper:3 outside:3 whose:4 encoded:1 heuristic:1 drawing:3 otherwise:1 erf:1 gi:1 statistic:2 think:1 highlighted:2 beal:1 product:4 zm:1 subgraph:1 flexibility:2 bug:7 moved:1 exploiting:2 parent:25 convergence:1 double:1 extending:1 produce:1 incremental:1 derive:1 illustrate:2 ac:3 develop:1 solves:1 implemented:2 involves:3 blanket:4 convention:1 filter:1 stochastic:1 generalization:2 summation:1 adjusted:1 extension:2 hold:2 lying:1 sufficiently:1 scope:1 substituting:1 adopt:1 early:1 purpose:2 currently:2 individually:1 correctness:1 successfully:2 tool:3 mit:2 gaussian:17 rather:1 ej:3 broader:1 jaakkola:1 publication:1 focus:1 modelling:2 likelihood:1 dragging:1 greatly:2 contrast:1 hsk:1 inference:9 initially:1 hidden:2 expressible:1 flexible:1 special:2 marginal:1 equal:2 construct:1 comprise:1 having:7 never:1 sampling:3 once:2 pai:5 broad:1 represents:1 bsu:2 xzj:1 few:1 gamma:5 divergence:2 pictorial:1 replaced:1 microsoft:2 interest:1 izj:2 message:4 weakness:1 mixture:9 diagnostics:1 chain:1 integral:1 necessary:1 experience:1 literally:1 logarithm:1 desired:1 increased:1 instance:2 earlier:1 lunn:1 applicability:1 front:1 iee:1 characterize:1 international:1 probabilistic:5 physic:1 together:3 iy:2 mouse:1 again:2 central:1 containing:1 choose:3 wishart:3 corner:1 account:1 coding:1 includes:1 explicitly:2 depends:1 script:2 view:2 performed:1 bayes:1 contribution:2 formed:2 qk:4 largely:2 efficiently:2 variance:1 correspond:1 bayesian:11 carlo:1 monitoring:1 mrc:3 rectified:1 j6:3 involved:1 associated:1 treatment:1 popular:1 logical:1 dimensionality:3 variationally:2 specify:1 yb:3 leen:2 evaluated:2 done:2 box:2 furthermore:2 just:1 stage:1 working:1 hand:5 christopher:2 replacing:1 propagation:1 logistic:1 indicated:2 building:1 effect:1 dietterich:1 concept:2 true:6 hence:1 leibler:2 neal:1 illustrated:1 during:1 noted:1 criterion:2 allowable:1 plate:5 complete:2 cp:2 interface:2 fj:1 variational:40 fi:3 common:1 sigmoid:1 multinomial:1 functional:2 stepping:1 volume:3 discussed:2 extend:2 approximates:1 interpretation:1 kluwer:2 cambridge:5 gibbs:3 dag:3 automatic:1 resorting:1 similarly:4 specification:1 compiled:1 gj:1 posterior:7 recent:1 success:1 cater:1 muller:1 seen:2 additional:1 somewhat:2 maximize:1 equally:1 qi:6 variant:1 basic:1 expectation:9 poisson:1 iteration:1 represent:1 achieved:1 whereas:1 conditionals:2 want:1 winn:2 interval:1 diagram:2 file:3 undirected:1 member:1 jordan:3 concerned:1 variety:3 xj:1 architecture:1 topology:1 simplifies:2 attias:1 qj:1 pca:4 effort:1 cause:1 dramatically:1 involve:1 characterise:1 eigenvectors:1 http:1 exist:1 zj:18 disjoint:1 write:1 discrete:9 group:2 key:3 monitor:1 drawn:2 cb3:1 graph:13 year:2 sum:3 run:1 package:2 inverse:2 powerful:1 place:1 family:6 extends:1 draw:1 bound:8 simplification:1 tackled:1 existed:1 strength:1 occur:1 software:1 generates:2 argument:1 winbugs:4 department:1 structured:1 combination:1 conjugate:8 slightly:1 em:1 suppressed:1 wi:4 making:1 restricted:1 recourse:1 ln:13 equation:11 computationally:2 conjugacy:4 rectification:1 turn:1 hh:1 tractable:6 end:1 adopted:1 available:1 operation:2 gaussians:1 apply:1 observe:2 hierarchical:1 probe:1 generic:2 appearing:2 alternative:1 shortly:1 florida:1 thomas:1 broaden:1 denotes:3 remaining:1 top:1 dirichlet:3 ensure:1 graphical:9 completed:1 bespoke:1 const:2 ghahramani:2 build:1 uj:3 approximating:1 disappear:1 unchanged:1 already:2 quantity:2 occurs:1 dependence:2 cycling:1 amongst:1 link:1 separate:1 code:4 kalman:1 length:1 relationship:2 expense:1 implementation:2 allowing:1 observation:2 revised:1 markov:6 finite:1 truncated:1 scripting:1 hinton:3 y1:1 ninth:1 arbitrary:4 hln:1 david:2 namely:2 required:2 specified:4 kl:4 optimized:2 pair:2 z1:2 engine:2 distinction:1 able:1 proceeds:1 challenge:1 memory:1 power:1 treated:1 natural:6 scheme:2 spiegelhalter:3 carried:1 categorical:1 coupled:1 tresp:1 review:1 prior:11 coloured:1 ksk:1 allocation:1 acyclic:2 switched:1 viewpoint:1 editor:4 prone:1 changed:1 keeping:1 side:6 allow:3 wide:4 saul:1 focussed:1 taking:1 sparse:1 distributed:1 calculated:1 dimension:1 valid:1 fb:1 wouldn:1 replicated:1 far:1 approximate:1 compact:1 alpha:1 kullback:2 confirm:1 global:1 consuming:1 factorize:1 xi:39 alternatively:1 continuous:4 latent:3 sk:6 icann:1 hierarchically:2 arrow:1 child:5 complementary:1 west:1 screen:2 precision:3 sub:2 wish:2 explicit:1 exponential:6 hw:1 down:1 bishop:3 specific:1 showing:2 explored:1 list:1 normalizing:1 intractable:1 intrinsic:1 ih:1 cpj:13 hui:2 conditioned:2 justifies:1 vibe:22 simply:3 conveniently:1 expressed:9 contained:1 ux:1 scalar:1 corresponds:3 satisfies:1 constantly:1 ma:2 conditional:10 goal:1 labelled:1 shared:2 determined:1 principal:2 called:4 accepted:1 evaluate:1 mcmc:2 |
1,288 | 2,173 | Incremental Gaussian Processes
?
Joaquin Quinonero-Candela
Informatics and Mathematical Modelling
Technical University of Denmark
DK-2800 Lyngby, Denmark
[email protected]
Ole Winther
Informatics and Mathematical Modelling
Technical University of Denmark
DK-2800 Lyngby, Denmark
[email protected]
Abstract
In this paper, we consider Tipping?s relevance vector machine (RVM)
[1] and formalize an incremental training strategy as a variant of the
expectation-maximization (EM) algorithm that we call Subspace EM
(SSEM). Working with a subset of active basis functions, the sparsity
of the RVM solution will ensure that the number of basis functions and
thereby the computational complexity is kept low. We also introduce
a mean field approach to the intractable classification model that is expected to give a very good approximation to exact Bayesian inference
and contains the Laplace approximation as a special case. We test the
algorithms on two large data sets with O(103 ? 104 ) examples. The results indicate that Bayesian learning of large data sets, e.g. the MNIST
database is realistic.
1 Introduction
Tipping?s relevance vector machine (RVM) both achieves a sparse solution like the support
vector machine (SVM) [2, 3] and the probabilistic predictions of Bayesian kernel machines
based upon a Gaussian process (GP) priors over functions [4, 5, 6, 7, 8]. Sparsity is interesting both with respect to fast training and predictions and ease of interpretation of the
solution. Probabilistic predictions are desirable because inference is most naturally formulated in terms of probability theory, i.e. we can manipulate probabilities through Bayes
theorem, reject uncertain predictions, etc.
It seems that Tipping?s relevance vector machine takes the best of both worlds. It is a GP
with a covariance matrix spanned by a small number of basis functions making the computational expensive matrix inversion operation go from O(N 3 ), where N is the number of
training examples to O(M 2 N ) (M being the number of basis functions). Simulation studies have shown very sparse solutions M N and good test performance [1]. However,
starting the RVM learning with as many basis functions as examples, i.e. one basis function
in each training input point, leads to the same complexity as for Gaussian processes (GP)
since in the initial step no basis functions are removed. That lead Tipping to suggest in
an appendix in Ref. [1] an incremental learning strategy that starts with only a single basis
function and adds basis functions along the iterations, and to formalize it very recently [9].
The total number of basis functions is kept low because basis functions are also removed.
In this paper we formalize this strategy using straightforward expectation-maximization
(EM) [10] arguments to prove that the scheme is the guaranteed convergence to a local
maximum of the likelihood of the model parameters.
Reducing the computational burden of Bayesian kernel learning is a subject of current
interest. This can be achieved by numerical approximations to matrix inversion [11] and
suboptimal projections onto finite subspaces of basis functions without having an explicit
parametric form of such basis functions [12, 13]. Using mixtures of GPs [14, 15] to make
the kernel function input dependent is also a promising technique. None of the Bayesian
methods can currently compete in terms of speed with the efficient SVM optimization
schemes that have been developed, see e.g. [3].
The rest of the paper is organized as follows: In section 2 we present the extended linear
models in a Bayesian perspective, the regression model and the standard EM approach.
In section 3, a variation of the EM algorithm, that we call the Subspace EM (SSEM) is
introduced that works well with sparse solution models. In section 4, we present the second
main contribution of the paper: a mean field approach to RVM classification. Section
5 gives results for the Mackey-Glass time-series and preliminary results on the MNIST
hand-written digits database. We conclude in section 6.
2 Regression
An extended linear model is built by transforming the input space by an arbitrary set of basis functions ?j : RD ? R that performs a non-linear transformation of the D-dimensional
input space. A linear model is applied to the transformed space whose dimension is equal
to the number of basis functions M :
y(xi ) =
M
X
j=1
?j ?j (xi ) = ?(xi ) ? ?
(1)
where ?(xi ) ? [?1 (xi ), . . . , ?M (xi )] denotes the ith row of the design matrix ? and ? =
(?1 , . . . , ?N )T is the weights vector. The output of the model is thus a linear superposition
of completely general basis functions. While it is possible to optimize the parameters of
the basis functions for the problem at hand [1, 16], we will in this paper assume that they
are given.
The simplest possible regression learning scenario can be described as follows: a set of
N input-target training pairs {xi , ti }N
i=1 are assumed to be independent and contaminated
with Gaussian noise of variance ? 2 . The likelihood of the parameters ? is given by
?N/2
1
? , ? 2 ) = 2?? 2
p(t|?
exp ? 2 kt ? ? ? k2
(2)
2?
where t = (t1 , . . . , tN )T is the target vector. Regularization is introduced in Bayesian
learning by means of a prior distribution over the weights. In general, the implied prior
over functions is a very complicated distribution. However, choosing a Gaussian prior on
the weights the prior over functions also becomes Gaussian, i.e. a Gaussian process. For
the specific choice of a factorized distribution with variance ??1
j :
r
1
?j
p(?j |?j ) =
exp ? ?j ?j2
(3)
2?
2
? ) is N (0, ?A?1 ?T ), i.e. a Gaussian process with covariance
the prior over functions p(y|?
function given by
M
X
1
?k (xi )?k (xj )
(4)
Cov(xi , xj ) =
?k
k=1
where ? = (?0 , . . . , ?N )T and A = diag(?0 , . . . , ?N ). We can now see how
sparseness in terms of the basis vectors may arise: if ??1
= 0 the kth basis vector
k
?k ? [?k (x1 ), . . . , ?k (xN )]T , i.e. the kth column in the design matrix, will not contribute
to the model. Associating a basis function with each input point may thus lead to a model
with a sparse representations in the inputs, i.e. the solution is only spanned by a subset of
all input points. This is exactly the idea behind the relevance vector machine, introduced
by Tipping [17]. We will see in the following how this also leads to a lower computational
complexity than using a regular Gaussian process kernel.
The posterior distribution over the weights?obtained through Bayes rule?is a Gaussian distribution
? , ? 2 )p(?
? |?
?)
p(t|?
? |t, ? , ? 2 ) =
? |?
?, ?)
p(?
= N (?
(5)
?, ? 2 )
p(t|?
?, ?) is a Gaussian distribution with mean ? and covariance ? evaluated at t.
where N (t|?
The mean and covariance are given by
? = ? ?2 ??T t
? = (? ?2 ?T ? + A)?1
(6)
(7)
The uncertainty about the optimal value of the weights captured by the posterior distribution (5) can be used to build probabilistic predictions. Given a new input x ? , the model
gives a Gaussian predictive distribution of the corresponding target t ?
Z
? |t, ? , ? 2 ) d?
? = N (t? |y? , ??2 )
p(t? |x? , ? , ? 2 ) = p(t? |x? , ? , ? 2 ) p(?
(8)
where
= ?(x? ) ? ?
y?
??2
2
= ? + ?(x? ) ? ? ? ?(x? )
(9)
T
(10)
For regression it is natural to use y? and ?? as the prediction and the error bar on the
prediction respectively. The computational complexity of making predictions is thus
O(M 2 P + M 3 + M 2 N ), where M is the number of selected basis functions (RVs) and P
is the number of predictions. The two last terms come from the computation of ? in eq.
(7).
The likelihood distribution over the training targets (2) can be ?marginalized? with respect
to the weights to obtain the marginal likelihood, which is also a Gaussian distribution
Z
? , ? 2 ) = p(t|?
? , ? 2 ) p(?
? |?
? ) d?
? = N (t|0, ? 2 I + ?A?1 ?T ) .
p(t|?
(11)
Estimating the hyperparameters {?j } and the noise ? 2 can be achieved by maximizing
(11). This is naturally carried out in the framework of the expectation-maximization (EM)
algorithm since the sufficient statistics of the weights (that act as hidden variables) are
available for this type of model. In other cases e.g. for adapting the length scale of the
kernel [4], gradient methods have to be used. For regression, the E-step is exact (the lower
bound on the marginal likelihood is made equal to the marginal likelihood) and consists in
estimating the mean and variance (6) and (7) of the posterior distribution of the weights
(5). For classification, the E-step will be approximate. In this paper we present a mean
field approach for obtaining the sufficient statistics.
The M-step corresponds to maximizing the expectation of the log marginal likelihood with
respect to the posterior, with respect to ? 2 and ? , which gives the following update
Prules:
1
1
1
2 new
2
2 old
=
,
and
(?
)
=
(||t
?
?
?||
+
(?
)
?new
=
2
2
j
j ?j ),
N
h? i
? +?jj
j
? |t,?
? ,? 2 )
p(?
j
where the quantity ?j ? 1 ? ?j ?jj is a measure of how ?well-determined? each weight ?j
is by the data [18, 1]. One can obtain a different update rule that gives faster convergence.
Although it is suboptimal in the EM sense, we have never observed it decrease the lower
bound on the marginal log-likelihood. The rule, derived in [1], is obtained by differentiation
of (11) and by an arbitrary choice of independent terms as is done by [18]. It makes use of
the terms {?j }:
?new
=
j
?j
?2j
(? 2 )new =
||t ? ? ?||2
P
.
N ? j ?j
(12)
In the optimization process many ?j grow to infinity, which effectively deletes the corresponding weight and basis function. Note that the EM update and the Mackay update
for ?j only implicitly depend upon the likelihood. This means that it is also valid for the
classification model we shall consider below.
A serious limitation of the EM algorithm and variants for problems of this type is that the
complexity of computing the covariance of the weights (7) in the E-step is O(M 3 +M 2 N ).
At least in the first iteration where no basis functions have been deleted M = N and we
are facing the same kind of complexity explosion that limits the applicability of Gaussian
processes to large training set. This has lead Tipping [1] to consider a constructive or
incremental training paradigm where one basis function is added before each E-step and
since basis functions are removed in the M-step, it turns out in practice that the total number
of basis functions and the complexity remain low [9]. In the following section we introduce
a new algorithm that formalizes this procedure that can be proven to increase the marginal
likelihood in each step.
3 Subspace EM
We introduce an incremental approach to the EM algorithm, the Subspace EM (SSEM), that
can be directly applied to training models like the RVM that rely on a linear superposition
of completely general basis functions, both for classification and for regression. Instead of
starting with a full model, i.e. where all the basis functions are present with finite ? values,
we start with a fully pruned model with all ?j set to infinity. Effectively, we start with no
model. The model is grown by iteratively including some ?j previously set to infinity to
the active set of ??s. The active set at iteration n, Rn , contains the indices of the basis
vectors with ? less than infinity:
R1 = 1
Rn = {i | i ? Rn?1 ? ?i ? L} ? {n}
(13)
where L is a finite very large number arbitrarily defined. Observe that R n contains at most
one more element (index) than Rn?1 . If some of the ??s indexed by Rn?1 happen to reach
L at the n-th step, Rn can contain less elements than Rn?1 . In figure 1 we give a schematic
description of the SSEM algorithm.
At iteration n the E-step is taken only in the subspace spanned by the weights whose
indexes are in Rn . This helps reducing the computational complexity of the M-step to
O(M 3 ), where M is the number of relevance vectors.
Since the initial value of ?j is infinity for all j, for regression the E-step yields always
an equality between the log marginal likelihood and its lower bound. At any step n, the
posterior can be exactly projected on to the space spanned by the weights ? j such that
j ? Rn , because the ?k = ? for all k not in Rn . Hence in the regression case, the SSEM
never decreases the log marginal likelihood. Figure 2 illustrates the convergence process
of the SSEM algorithm compared to that of the EM algorithm for regression.
Set ?j = L for all j. (L is a very large number) Set n = 1
Update the set of active indexes Rn
Perform an E-step in subspace ?j such that j ? Rn
Perform the M-step for all ?j such that j ? Rn
If visited all basis functions, end, else go to 2.
1.
2.
3.
4.
5.
Figure 1: Schematics of the SSEM algorithm.
Number of RVs vs. CPU time
Likelihood vs. CPU time
450
standard EM
SSEM
1200
400
1000
350
300
Number of RVs
Log marginal likelihood
800
600
400
200
250
200
150
0
100
?200
50
SSEM
standard EM
?400
0
20
40
60
80
CPU time (seconds)
100
120
0
0
20
40
60
80
CPU time (seconds)
100
120
Figure 2: Training on 400 samples of the Mackey-Glass time series, testing on 2000 cases.
Log marginal likelihood as a function of the elapsed CPU time (left) and corresponding
number of relevance vectors (right) for both SSEM and EM.
We perform one EM step for each time a new basis function is added to the active set. Once
all the examples have been visited, we switch to the batch EM algorithm on the active set
until some convergence criteria has been satisfied, for example until the relative increase in
the likelihood is smaller than a certain threshold. In practice some 50 of these batch EM
iterations are enough.
4 Classification
Unlike the model discussed above, analytical inference is not possible for classification
models. Here, we will discuss the adaptive TAP mean field approach?initially proposed for
Gaussian processes [8]?that are readily translated to RVMs. The mean field approach has
the appealing features that it retains the computational efficiency of RVMs, is exact for the
regression and reduces to the Laplace approximation in the limit where all the variability
comes from the prior distribution.
We consider binary t = ?1 classification using the probit likelihood with ?input? noise ? 2
y(x)
,
(14)
p(t|y(x)) = erf t
?
?
Rx
2
where Dz ? e?z /2 dz/ 2? and erf(x) ? ?? Dz is an error function (or cumulative
Gaussian distribution). The advantage of using this sigmoid rather than the commonly
used 0/1-logistic is that we under the mean field approximation
can derive an analytical
R
expression for the predictive distribution p(t? |x? , t) = p(t? |y)p(y|x? , t)dy needed for
making Bayesian predictions. Both a variational and the advanced mean field approach?
used here?make a Gaussian approximation for p(y|x? , t) [8] with mean and variance given
by regression results y? and ??2 ? ?
? 2 , and y? and ??2 given by eqs. (9) and (10). This leads
to the following approximation for the predictive distribution
Z
y
y?
p(y|x? , t) dy = erf t?
.
p(t? |x? , t) =
erf t?
?
??
(15)
However, the mean and covariance of the weights are no longer found by analytical expressions, but has to be obtained from a set of non-linear mean field equations that also follow
from equivalent assumptions of Gaussianity for the training set outputs y(x i ) in averages
over reduced (or cavity) posterior averages.
In the following, we will only state the results which follows from combining the RVM
Gaussian process kernel (4) with the results of [8]. The sufficient statistics of the weights
are written in terms of a set of O(N ) mean field parameters
? = A?1 ?T ?
T
? =
where ?i ?
?
?yic
Z(yic , Vic
A + ? ??
ln Z(yic , Vic + ? 2 ) and
2
+? )
?
Z
p(ti |yic
(16)
?1
q
+ z Vic + ? 2 ) Dz = erf
(17)
yc
ti p c i
Vi + ? 2
!
. (18)
The last equality holds for the likelihood eq. (14) and yic and Vic are the mean and variance
of the so called cavity field. The mean value is yic = ?(xi ) ? ? ? Vic ?i . The distinction
c
c
between the different approximation
schemes
is solely in the variance V i : Vi = 0 is the
c
?1 T
Laplace approximation, Vi = ?A ? ii is the so called naive mean field theory and
an improved estimate is available from the adaptive TAP mean field theory [8]. Lastly, the
diagonal matrix ? is the equivalent of the noise variance in the regression model (compare
??i
c ??i
eqs. (17) and (7) and is given by ?i = ? ?y
c /(1+Vi ?y c ) . This set of non-linear equations
i
i
are readily solved (i.e. fast and stable) by making Newton-Raphson updates in ? treating
the remaining quantities as help variables:
? = (I + A?1 ?T ??)?1 (A?1 ?T ? ? ? ) = ?(?T ? ? A?
?)
??
(19)
The computational complexity of the E-step for classification is augmented with respect to
the regression case by the fact that it is necessary to construct and invert a M ? M matrix
usually many times (typically 20), once for each step of the iterative Newton method.
5 Simulations
We illustrate the performance of the SSEM for regression on the Mackey-Glass chaotic
time series, which is well-known for its strong non-linearity. In [16] we showed that the
RVM has an order of magnitude superior performance than carefully tuned neural networks
for time series prediction on the Mackey-Glass series. The inputs are formed by L = 16
samples spaced 6 periods from each other xk = [z(k ? 6), z(k ? 12), . . . , z(k ? 6L)] and
the targets are chosen to be tk = z(k) to perform six steps ahead prediction (see [19] for
details). We use Gaussian basis functions of fixed variance ? 2 = 10. The test set comprises
5804 examples.
We perform prediction experiments for different sizes of the training set. We perform in
each case 10 repetitions with different partitions of the data sets into training and test. We
compare the test error, the number of RVs selected and the computer time needed for the
batch and the SSEM method. We present the results obtained with the growth method
relative to the results obtained with the batch method in figure 3. As expected, the relative
Mackey?Glass data
3
growth
Classification on MNIST digits
batch
Ete
/Ete
Tcpugrowth/Tcpubatch
NRVgrowth/NRVbatch
2.5
1.5
1
2
1.5
0.5
1
0
Training error prob.
Test error prob.
Scaled loglik
0.5
0
0
500
1000
1500
2000
Number of training examples
?0.5
0
200
400
Iteration
600
800
Figure 3: Left: Regression, mean values over 10 repetitions of relative test error, number
of RVs and computer time for the Mackey-Glass data, up to 2400 training examples and
5804 test examples. Right: Classification, Log marginal likelihood, test and training errors
while training on one class against all the others, 60000 training and 10000 test examples.
computer time of the growth method compared with the batch method decreases with size
of the training set. For a few thousand examples the SSEM method is an order of magnitude
faster than the batch method. The batch method proved only to be faster for 100 training
examples, and could not be used with data sets of thousands of examples on the machine on
which we run the experiments because of its high memory requirements. This is the reason
why we only ran the comparison for up to 2400 training example for the Mackey-Glass
data set.
Our experiments for classification are at the time of sending this paper to press very premature: we choose a very large data set, the MNIST database of handwritten digits [20],
with 60000 training and 10000 test images. The images are of size 28 ? 28 pixels. We
use PCA to project them down to 16 dimensional vectors. We only perform a preliminary
experiment consisting of a one against all binary classification problem to illustrate that
Bayesian approaches to classification can be used on very large data sets with the SSEM
algorithm. We train on 13484 examples (the 6742 one?s and another 6742 random non-one
digits selected at random from the rest) and we use 800 basis functions for both the batch
and Subspace EM. In figure 3 we show the convergence of the SSEM in terms of the log
marginal likelihood and the training and test probabilities of error. The test probability of
error we obtain is 0.74 percent with the SSEM algorithm and 0.66 percent with the batch
EM. Under the same conditions the SSEM needed 55 minutes to do the job, while the batch
EM needed 186 minutes. The SSEM gives a machine with 28 basis functions and the batch
EM one with 31 basis functions.
6 Conclusion
We have presented a new approach to Bayesian training of linear models, based on a subspace extension of the EM algorithm that we call Subspace EM (SSEM). The new method
iteratively builds models from a potentially big library of basis functions. It is especially
well-suited for models that are constructed such that they yield a sparse solution, i.e. the solution is spanned by small number M of basis functions, which is much smaller than N , the
number of examples. A prime example of this is Tipping?s relevance vector machine that
typically produces solutions that are sparser than those of support vector machines. With
the SSEM algorithm the computational complexity and memory requirement decrease from
O(N 3 ) and O(N 2 ) to O(M 2 N ) (somewhat higher for classification) and O(N M ). For
classification, we have presented a mean field approach that is expected to be a very good
approximation to the exact inference and contains the widely used Laplace approximation
as an extreme case. We have applied the SSEM algorithm to both a large regression and a
large classification data sets. Although preliminary for the latter, we believe that the results
demonstrate that Bayesian learning is possible for very large data sets. Similar methods
should also be applicable beyond supervised learning.
Acknowledgments JQC is funded by the EU Multi-Agent Control Research Training Network - EC
TMR grant HPRNCT-1999-00107. We thank Lars Kai Hansen for very useful discussions.
References
[1] Michael E. Tipping, ?Sparse bayesian learning and the relevance vector machine,? Journal of
Machine Learning Research, vol. 1, pp. 211?244, 2001.
[2] Vladimir N. Vapnik, Statistical Learning Theory, Wiley, New York, 1998.
[3] Bernhard Sch?olkopf and Alex J. Smola, Learning with Kernels, MIT Press, Cambridge, 2002.
[4] Carl E. Rasmussen, Evaluation of Gaussian Processes and Other Methods for Non-linear
Regression, Ph.D. thesis, Dept. of Computer Science, University of Toronto, 1996.
[5] Chris K. I. Williams and Carl E. Rasmussen, ?Gaussian Proceses for Regression,? in Advances
in Neural Information Processing Systems, 1996, number 8, pp. 514?520.
[6] D. J. C. Mackay, ?Gaussian Processes: A replacement for supervised Neural Networks?,? Tech.
Rep., Cavendish Laboratory, Cambridge University, 1997, Notes for a tutorial at NIPS 1997.
[7] Radford M. Neal, Bayesian Learning for Neural Networks, Springer, New York, 1996.
[8] Manfred Opper and Ole Winther, ?Gaussian processes for classification: Mean field algorithms,? Neural Computation, vol. 12, pp. 2655?2684, 2000.
[9] Michael Tipping and Anita Faul, ?Fast marginal likelihood maximisation for sparse bayesian
models,? in International Workshop on Artificial Intelligence and Statistics, 2003.
[10] N. M. Dempster, A.P. Laird, and D. B. Rubin, ?Maximum likelihood from incomplete data via
the EM algorithm,? J. R. Statist. Soc. B, vol. 39, pp. 185?197, 1977.
[11] Chris Williams and Mathias Seeger, ?Using the Nystr?om method to speed up kernel machines,?
in Advances in Neural Information Processing Systems, 2001, number 13, pp. 682?688.
[12] Alex J. Smola and Peter L. Bartlett, ?Sparse greedy gaussian process regression,? in Advances
in Neural Information Processing Systems, 2001, number 13, pp. 619?625.
[13] Lehel Csat?o and Manfred Opper, ?Sparse representation for gaussian process models,? in
Advances in Neural Information Processing Systems, 2001, number 13, pp. 444?450.
[14] Volker Tresp, ?Mixtures of gaussian processes,? in Advances in Neural Information Processing
Systems, 2000, number 12, pp. 654?660.
[15] Carl E. Rasmussen and Zoubin Ghahramani, ?Infinite mixtures of gaussian process experts,? in
Advances in Neural Information Processing Systems, 2002, number 14.
[16] Joaquin Qui?nonero-Candela and Lars Kai Hansen, ?Time series prediction based on the relevance vector machine with adaptive kernels,? in International Conference on Acoustics, Speech,
and Signal Processing (ICASSP), 2002.
[17] Michael E. Tipping, ?The relevance vector machine,? in Advances in Neural Information
Processing Systems, 2000, number 12, pp. 652?658.
[18] David J. C. MacKay, ?Bayesian interpolation,? Neural Computation, vol. 4, no. 3, pp. 415?447,
1992.
[19] Claus Svarer, Lars K. Hansen, Jan Larsen, and Carl E. Rasmussen, ?Designer networks for time
series processing,? in IEEE NNSP Workshop, 1993, pp. 78?87.
[20] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, ?Gradient-based learning applied to document
recognition,? in Poceedings of the IEEE, 1998, vol. 86, pp. 2278?2324.
| 2173 |@word inversion:2 seems:1 simulation:2 covariance:6 thereby:1 nystr:1 initial:2 contains:4 series:7 tuned:1 document:1 current:1 loglik:1 written:2 readily:2 numerical:1 partition:1 realistic:1 happen:1 treating:1 update:6 mackey:7 v:2 intelligence:1 selected:3 greedy:1 xk:1 nnsp:1 ith:1 manfred:2 contribute:1 toronto:1 mathematical:2 along:1 constructed:1 prove:1 consists:1 introduce:3 expected:3 multi:1 cpu:5 becomes:1 project:1 estimating:2 linearity:1 factorized:1 kind:1 developed:1 transformation:1 differentiation:1 formalizes:1 ti:3 act:1 growth:3 exactly:2 k2:1 scaled:1 control:1 grant:1 rvms:2 t1:1 before:1 local:1 limit:2 solely:1 interpolation:1 ease:1 acknowledgment:1 lecun:1 testing:1 practice:2 maximisation:1 chaotic:1 digit:4 procedure:1 jan:1 reject:1 adapting:1 projection:1 regular:1 suggest:1 zoubin:1 onto:1 optimize:1 equivalent:2 dz:4 maximizing:2 go:2 straightforward:1 starting:2 williams:2 rule:3 spanned:5 cavendish:1 variation:1 laplace:4 target:5 exact:4 gps:1 carl:4 element:2 expensive:1 recognition:1 database:3 observed:1 solved:1 thousand:2 eu:1 decrease:4 removed:3 ran:1 transforming:1 dempster:1 complexity:10 depend:1 predictive:3 upon:2 efficiency:1 basis:37 completely:2 translated:1 icassp:1 grown:1 train:1 fast:3 ole:2 artificial:1 choosing:1 whose:2 widely:1 kai:2 cov:1 statistic:4 erf:5 gp:3 laird:1 advantage:1 analytical:3 j2:1 combining:1 nonero:1 description:1 olkopf:1 convergence:5 requirement:2 r1:1 produce:1 incremental:5 tk:1 help:2 derive:1 illustrate:2 job:1 eq:4 soc:1 strong:1 proceses:1 faul:1 indicate:1 come:2 lars:3 preliminary:3 extension:1 hold:1 exp:2 achieves:1 applicable:1 currently:1 superposition:2 visited:2 rvm:8 hansen:3 repetition:2 mit:1 gaussian:27 always:1 rather:1 volker:1 derived:1 modelling:2 likelihood:22 tech:1 seeger:1 sense:1 glass:7 inference:4 dependent:1 anita:1 typically:2 lehel:1 initially:1 hidden:1 transformed:1 pixel:1 classification:18 special:1 mackay:3 marginal:13 field:14 equal:2 never:2 having:1 once:2 construct:1 contaminated:1 others:1 serious:1 few:1 ete:2 consisting:1 replacement:1 interest:1 evaluation:1 mixture:3 extreme:1 behind:1 kt:1 explosion:1 necessary:1 indexed:1 incomplete:1 old:1 uncertain:1 column:1 retains:1 maximization:3 applicability:1 subset:2 winther:2 international:2 probabilistic:3 informatics:2 michael:3 tmr:1 jqc:2 thesis:1 satisfied:1 choose:1 expert:1 gaussianity:1 vi:4 candela:2 start:3 bayes:2 complicated:1 contribution:1 om:1 formed:1 variance:8 yield:2 spaced:1 bayesian:15 handwritten:1 none:1 rx:1 reach:1 against:2 pp:12 larsen:1 naturally:2 proved:1 organized:1 formalize:3 carefully:1 higher:1 tipping:10 supervised:2 follow:1 improved:1 evaluated:1 done:1 smola:2 lastly:1 until:2 joaquin:2 working:1 hand:2 logistic:1 believe:1 contain:1 regularization:1 equality:2 hence:1 iteratively:2 laboratory:1 neal:1 criterion:1 demonstrate:1 tn:1 performs:1 percent:2 image:2 variational:1 recently:1 sigmoid:1 superior:1 discussed:1 interpretation:1 cambridge:2 rd:1 funded:1 stable:1 longer:1 etc:1 add:1 posterior:6 showed:1 perspective:1 prime:1 scenario:1 certain:1 binary:2 arbitrarily:1 rep:1 captured:1 somewhat:1 paradigm:1 period:1 signal:1 ii:1 rv:5 full:1 desirable:1 reduces:1 technical:2 faster:3 raphson:1 manipulate:1 schematic:2 prediction:14 variant:2 regression:19 expectation:4 iteration:6 kernel:9 achieved:2 invert:1 else:1 grow:1 sch:1 rest:2 unlike:1 claus:1 subject:1 call:3 bengio:1 enough:1 switch:1 xj:2 associating:1 suboptimal:2 idea:1 haffner:1 expression:2 six:1 pca:1 bartlett:1 peter:1 speech:1 york:2 jj:2 useful:1 ph:1 statist:1 simplest:1 reduced:1 tutorial:1 designer:1 csat:1 shall:1 vol:5 threshold:1 deletes:1 deleted:1 kept:2 compete:1 prob:2 run:1 uncertainty:1 appendix:1 dy:2 qui:1 bound:3 guaranteed:1 ahead:1 infinity:5 alex:2 speed:2 argument:1 pruned:1 remain:1 smaller:2 em:27 appealing:1 making:4 taken:1 lyngby:2 equation:2 ln:1 previously:1 turn:1 discus:1 needed:4 end:1 sending:1 available:2 operation:1 observe:1 batch:12 denotes:1 remaining:1 ensure:1 marginalized:1 newton:2 ghahramani:1 build:2 especially:1 implied:1 added:2 quantity:2 strategy:3 parametric:1 diagonal:1 gradient:2 kth:2 subspace:10 thank:1 quinonero:1 chris:2 reason:1 denmark:4 length:1 index:4 vladimir:1 potentially:1 design:2 perform:7 finite:3 extended:2 variability:1 rn:13 arbitrary:2 introduced:3 david:1 pair:1 tap:2 acoustic:1 elapsed:1 distinction:1 nip:1 beyond:1 bar:1 below:1 usually:1 yc:1 sparsity:2 built:1 including:1 memory:2 natural:1 rely:1 advanced:1 scheme:3 vic:5 library:1 dtu:2 carried:1 naive:1 tresp:1 prior:7 relative:4 fully:1 probit:1 interesting:1 limitation:1 proven:1 facing:1 agent:1 sufficient:3 rubin:1 row:1 last:2 rasmussen:4 sparse:9 dimension:1 xn:1 world:1 valid:1 cumulative:1 opper:2 made:1 adaptive:3 projected:1 commonly:1 premature:1 ec:1 approximate:1 implicitly:1 bernhard:1 cavity:2 imm:2 active:6 conclude:1 assumed:1 xi:10 iterative:1 why:1 yic:6 promising:1 obtaining:1 bottou:1 diag:1 main:1 owi:1 big:1 noise:4 arise:1 hyperparameters:1 ref:1 x1:1 augmented:1 wiley:1 comprises:1 explicit:1 theorem:1 down:1 minute:2 specific:1 dk:4 svm:2 intractable:1 burden:1 mnist:4 vapnik:1 workshop:2 effectively:2 magnitude:2 illustrates:1 sparseness:1 sparser:1 suited:1 radford:1 springer:1 corresponds:1 formulated:1 determined:1 infinite:1 reducing:2 total:2 called:2 mathias:1 svarer:1 support:2 latter:1 relevance:10 constructive:1 dept:1 |
1,289 | 2,174 | Application of Variational Bayesian Approach to
Speech Recognition
Shinji Watanabe, Yasuhiro Minami, Atsushi Nakamura and Naonori Ueda
NTT Communication Science Laboratories, NTT Corporation
2-4, Hikaridai, Seika-cho, Soraku-gun, Kyoto, Japan
{watanabe,minami,ats,ueda}@cslab.kecl.ntt.co.jp
Abstract
In this paper, we propose a Bayesian framework, which constructs
shared-state triphone HMMs based on a variational Bayesian approach,
and recognizes speech based on the Bayesian prediction classification;
variational Bayesian estimation and clustering for speech recognition
(VBEC). An appropriate model structure with high recognition performance can be found within a VBEC framework. Unlike conventional
methods, including BIC or MDL criterion based on the maximum likelihood approach, the proposed model selection is valid in principle, even
when there are insufficient amounts of data, because it does not use
an asymptotic assumption. In isolated word recognition experiments,
we show the advantage of VBEC over conventional methods, especially
when dealing with small amounts of data.
1 Introduction
A statistical modeling of spectral features of speech (acoustic modeling) is one of the most
crucial parts in the speech recognition. In acoustic modeling, a triphone-based hidden
Markov model (triphone HMM) has been widely employed. The triphone is a context
dependent phoneme unit that considers both the preceding and following phonemes. Although the triphone enables the precise modeling of spectral features, the total number of
triphones is too large to prepare sufficient amounts of training data for each triphone. In
order to deal with the problem of data insufficiency, an HMM state is usually shared among
multiple triphone HMMs, which means the amount of training data per state inflates. Such
shared-state triphone HMMs (SST-HMMs) can be constructed by successively clustering
states based on the phonetic decision tree method [4] [7]. The important practical problem
that must be solved when constructing SST-HMMs is how to optimize the total number
of shared states adaptively to the amounts of available training data. Namely, maintaining
the balance between model complexity and training data size is quite important for high
generalization performance.
The maximum likelihood (ML) is inappropriate as a model selection criterion since ML
increases monotonically as the number of states increases. Some heuristic thresholding
is therefore necessary to terminate the partitioning. To solve this problem, the Bayesian
information criterion (BIC) and minimum description length (MDL) criterion have been
employed to determine the tree structure of SST-HMMs [2] [5] 1 . However, since the
BIC/MDL is based on an asymptotic assumption, it is invalid in principle when the number
of training data is small because of the failure of the assumption.
In this paper, we present a practical method within the Bayesian framework for estimating posterior distributions over parameters and selecting an appropriate model structure of
SST-HMMs (clustering triphone HMM states) based on a variational Bayesian (VB) approach, and recognizing speech based on the Bayesian prediction classification: variational
Bayesian estimation and clustering for speech recognition (VBEC). Unlike the BIC/MDL,
VB does not assume asymptotic normality, and it is therefore applicable in principle, even
when there are insufficient data. The VB approach has been successfully applied to model
selection problems, but mainly for relatively simple mixture models [1] [3] [6] [8]. Here,
we try to apply VB to SST-HMMs with more a complex model structure than the mixture
model and evaluate the effectiveness through a large-scale real speech recognition experiment.
2 Variational Bayesian framework
First, we briefly review the VB framework. Let O be a given data set. In the Bayesian
approach we are interested in posterior distributions over model parameters, p(?|O, m),
and the model structure, p(m|O). Here, ? is a set of model parameters and m is an index
of the model structure. Let us consider a general probabilistic model with latent variables.
Let Z be a set of latent variables. Then the model with a fixed model structure m can be
defined by the joint distribution p(O, Z|?, m).
In VB, variational posteriors q(?|O, m), q(Z|O, m), and q(m|O) are introduced to approximate the true corresponding posteriors. The optimal variational posteriors over ? and
Z, and the appropriate model structure that maximizes the optimal q(m|O) can be obtained
by maximizing the following objective function:
p(O, Z|?, m)p(?|m)
,
(1)
Fm [q] =
log
q(Z|O, m)q(?|O, m) q(Z|O,m),q(? |O,m)
w.r.t. q(?|O, m), q(Z|O, m), and m. Here hf (x)ip(x) denotes the expectation of f (x)
w.r.t. p(x). p(?|m) is a prior distribution. This optimization can be effectively performed
by an EM-like iterative algorithm (see [1] for the details).
3 Applying a VB approach to acoustic models
3.1
Output distributions and prior distributions
We attempt to apply a VB approach to a left-to-right HMM, which has been widely used
to represent a phoneme unit in acoustic models for speech recognition, as shown in Figure
1. Let O = {O t ? RD : t = 1, ..., T } be a sequential data set for a phoneme unit. The
output distribution in an HMM is given by
YT
p(O, S, V |?, m) =
ast?1 st cst vt bst vt (O t ),
(2)
t= 1
where S is a set of sequences of hidden states, V is a set of sequences of Gaussian mixture
components, and st and v t denote the state and mixture components at time t. S and V are
sets of discrete latent variables that correspond to Z mentioned above. aij denotes the state
1
These criteria have been independently proposed, but they are practically the same. Therefore,
we refer to them hereafter as BIC/MDL.
a11
i =1
a33
a22
a12
i=2
Figure 1: Hidden Markov model for each
phoneme unit. A state is represented by
the Gaussian mixture distribution below
the state. There are three states and three
Gaussian components in this figure.
i=3
a 23
Gaussian mixture for state i
transition probability from state i to state j, and cjk is the k-th weight factor of the Gaussian
mixture for state j. bjk (= N (O t |?jk , ?jk )) denotes the Gaussian distribution with mean
vector ?jk and covariance ?jk . ? = {aij , cjk , ?jk , ??1
jk |i, j = 1, ..., J, k = 1, ..., L} is
a set of model parameters. J denotes the number of states in an HMM and L denotes the
number of Gaussian components in a state. In this paper, we restrict covariance matrices in
the Gaussian distribution to diagonal ones. The conjugate prior distributions are assumed
to be as follows:
Y
0
p(?|m) =
D({aij0 }Jj0 = 1 |?0 )D({cjk0 }L
k0 = 1 |? )
i,j,k
YD
0
0
? N (?jk |? 0jk , (? 0 )?1 ?jk )
G(??1
(3)
jk,d |? , Rjk,d ).
d= 1
0
?0 = {?0 , ?0 , ? 0jk , ? 0 , ? 0 , Rjk
} is a set of hyperparameters. We assume the hyperparameters are constants. In Eq.(3), D denotes a Dirichlet distribution and G denotes a gamma
distribution.
3.2
Optimal variational posterior distribution q?(?|O, m)
From the output distributions and prior distributions in section 3.1, the optimal variational
posterior distribution q?(?|O, m) can be obtained as:
D({aij }Jj= 1 |{??ij }Jj= 1 )
D({cjk }L
?jk }L
)
k= 1 |{?
k= 1
QD
?1
? jk,d ),
?jk , R
N (?jk |?
? jk , ??jk ?jk ) d= 1 G(??1
jk,d |?
q?({aij }Jj= 1 |O, m) =
q?({cjk }L
k= 1 |O, m) =
q?(bjk |O, m)
=
(4)
? ?,
? ??, R
? ? {?,
? jk } is a set of posterior distribution parameters defined as:
? jk , ?,
?
? ?
XT
? jk = ? 0 ? 0jk +
??ij = ?0 + ??ij , ??jk = ?0 + ??jk , ??jk = ? 0 + ??jk , ?
??t O t /??jk ,
t= 1 jk
XT
t
? jk,d = R0 + ? 0 (? 0 ? ??jk,d )2 +
??jk = ? 0 + ??jk , R
??jk
(Odt ? ??jk,d )2 .
(5)
jk,d
jk,d
t= 1
t ?t
? is composed of
, ?jk ? q?(st = j, v t =
? q?(s = i, s
= j|O, m), ??ij ? ?Tt= 1 ??ij
?
t
t
k|O, m) and ??jk ? ?Tt= 1 ??jk
. ??ij
denotes the transition probability from state i to state j at
t
?
time t. ?jk denotes the occupation probability on mixture component k in state j at time t.
t
??ij
3.3
t
t+ 1
Optimal variational posterior distribution q?(S, V |O, m)
From the output distributions and prior distributions in section 3.1, the optimal variational
posterior distribution over latent variables q?(S, V |O, m) can be obtained as:
YT
q?(S, V |O, m) ?
a
?st?1 st c?st vt ?bst vt (Ot ),
(6)
t= 1
where
a
?st?1 st
=
c?st vt
=
?bst vt (Ot ) =
XJ
e x p ? (??st?1 st ) ? ? (
?? t?1 st0 ) ,
st0 =1 s
XL
e x p ? (??st vt ) ? ? (
?? t t0 ) ,
v t0 =1 s v
?
t
t
t
t
?s v /2) ?
e x p D/2 log 2? ? 1/?s v + ? (?
XD
? st vt ,d /2) + (Ot ? ??st vt ,d )2 ??st vt /R
? st vt ,d
log(R
. (7)
?1/2
d
d=1
t
? (y) is a digamma function. From these results, transition and occupation probability ??ij
t
and ??ij can be obtained by using either a deterministic assignment via the Viterbi algorithm
or a probabilistic assignment via the Forward-Backward algorithm. Thus, q?(? |O, m) and
q?(S, V |O, m) can be calculated iteratively that result in maximizing Fm .
4 VB training algorithm for acoustic models
Based on the discussion in section 3, a VB training algorithm for an acoustic model based
on an HMM and Gaussian mixture model with a fixed model structure m is as follows:
????????????????????????????????????
t
t
Step 1) Initialize ??ij
[? = 0], ??ij
[? = 0] and set ?0 .
Step 2) Compute q(S, V |O, m)[? + 1] using ?? t [? ], ??t [? ] and ?0 .
ij
t
t
?
Update ??ij [? +1] and ?ij [? +1] using q(S, V
ij
|O, m)[? +1] via the Viterbi algorithm
or Forward-Backward algorithm.
? + 1] using ?? t [? + 1], ??t [? + 1] and ?0 .
Step 4) Compute ?[?
ij
ij
? + 1] and calculate Fm [? ] based on
Step 5) Compute q(? |O, m)[? + 1] using ?[?
q(? |O, m)[? + 1] and q(S, V |O, m)[? + 1].
Step 6) If |(Fm [? + 1] ? Fm [? ])/Fm [? + 1]| ? ?, then stop; otherwise set ? ? ? + 1 and
go to Step 2.
????????????????????????????????????
? denotes an iteration count. In our experiments, we employed the Viterbi algorithm in
Step 3.
Step 3)
5 Variational Bayesian estimation and clustering for speech
recognition
In the previous section, we described a VB training algorithm for HMMs. Here, we explain
VBEC, which constructs an acoustic model based on SST-HMMs and recognizes speech
based on the Bayesian prediction classification. VBEC consists of three phases: model
structure selection, retraining and recognition. The model structure is determined based on
triphone-state clustering by using the phonetic decision tree method [4] [7]. The phonetic
decision tree is a kind of binary tree that has a phonetic ?Yes/No? question attached at each
node, as shown in Figure 2. Let ? (n) denote a set of states held by a tree node n. We
start with only a root node (n = 0), which holds a set of all the triphone HMM states
? (0) for an identical center phoneme. The set of triphone states is then split into two sets,
? (nY ) and ? (nN ), which are held by two new nodes, nY and nN , respectively, as shown
in Figure 3. The partition is determined by an answer to a phonetic question such as ?is
the preceding phoneme a vowel?? or ?is the following phoneme a nasal?? We choose a
particular question for a node that maximize the gain of F m when the node is split into two
*/a(i)/*
Yes
Yes
No
No
Yes
k/a(i)/i
k/a(i)/o
Yes
Yes
?(n)
No
Yes
No
n
No
Phonetic
question
No
root n od e(n=0)
?
ts/a(i)/m
ch/a(i)/n g
nY
?(nY)
nN
?(nN)
leaf n od e
?
Figure 2: A set of all triphone HMM
states */a(i)/* is clustered based on the
phonetic decision tree method.
Figure 3: Splitting a set of triphone
HMM states ?(n) into two sets ?(nY )
?(nN ) by answering phonetic questions
according to an objective function.
nodes, and if all the questions decrease F m after splitting, we stop splitting. We continue
this splitting successively for every new set of states to obtain a binary tree, each leaf node
of which holds a clustered set of triphone states. The states belonging to the same cluster
are merged into a single state. A set of triphones is thus represented by a set of sharedstate triphone HMMs (SST-HMMs). An HMM, which represents a phonemic unit, usually
consists of a linear sequence of three or four states. A decision tree is produced specifically
for each state in the sequence, and the trees are independent of each other.
Note that in the triphone-states clustering mentioned above, we assume the following conditions to reduce computations:
? The state assignments while splitting are fixed.
? A single Gaussian distribution for one state is used.
? Contributions of the transition probabilities to the objective function are ignored.
By using these conditions, latent variables are removed. As a result, all variational posteriors and Fm can be obtained as closed forms without an iterative procedure.
Once we have obtained the model structure, we retrain the posterior distributions using the VB algorithm given in section 4. In recognition, an unknown datum xt for a
frame t is classified as the optimal phoneme class y using the predictive posterior clast
sification probability p(y|xt , O, m)
? ? p(y)p(xt |y, O, m)/p(x
?
) for the estimated model
structure m.
? Here, p(y) is the class prior obtained by language and lexcon models, and
p(xt |y, O, m)
? is the predictive density. If we approximate the true posterior p(?|y, O, m)
?
by the estimated variational
posteriors q?(?|y, O, m),
? p(xt |y, O, m)
? can be calculated by
R
p(xt |y, O, m)
? ? p(xt |y, ?, m)?
? q (?|y, O, m)d?.
?
Therefore, the optimal class y can be
obtained by
Z
0 t
0
y = arg m ax
p(y |x , O, m)
? ? arg m ax
p(y ) p(xt |y 0 , ?, m)?
? q (?|y, O, m)d?.
?
(8)
0
0
y
y
In the calculation of (8), the integral over Gaussian means and covariances for a frame can
be solved analytically to be Student distributions. Therefore, we can compute a Bayesian
predictive score for a frame, and then can compute a phoneme sequence score by using
the Viterbi algorithm. Thus, we can construct a VBEC framework for speech recognition
by selecting an appropriate model structure and estimating posterior distributions with the
VB approach, and then obtaining a recognition result based on the Bayesian prediction
classification.
Table 1: Acoustic conditions
Sampling rate
16 kHz
Quantization
16 bit
12 - order MFCC
Feature vector
with ? MFCC
Hamming
Window
Frame size/shift 25/10 ms
Table 2: Prepared HMM
# of states
# of phoneme categories
Output distribution
3 (Left to right)
27
Single Gaussian
6 Experiments
We conducted two experiments to evaluate the effectiveness of VBEC. The first experiment compared VBEC with the conventional ML-BIC/MDL method for variable amounts
of training data. In the ML-BIC/MDL, retraining and recognition are based on the ML
approach and model structure selection is based on the BIC/MDL. The second experiment
examined the robustness of the recognition performance with preset hyperparameter values
against changes in the amounts of training data.
6.1
VBEC versus ML-BIC/MDL
The experimental conditions are summarized in Tables 1 and 2. As regards the hyperparameters, the mean and variance values of the Gaussian distribution were set at ? 0 and
R0 in each root node, respectively, and the heuristics were removed for ? 0 and R0 . The
determination of ? 0 and ? 0 was still heuristic. We set ? 0 = ? 0 = 0.01, each of which were
determined experimentally. The training and recognition data used in these experiments
are shown in Table 3.
The total training data consisted of about 3,000 Japanese sentences spoken by 30 males.
These sentences were designed so that the phonemic balance was maintained. The total
recognition data consisted of 2,500 Japanese city names spoken by 25 males. Several
subsets were randomly extracted from the training data set, and each subset was used to
construct a set of SST-HMMs. As a result, 40 sets of SST-HMMs were prepared for several
subsets of training data.
Figures 4 and 5 show the recognition rate and the total number of states in a set of SSTHMMs, according to the varying amounts of training data. As shown in Figure 4, when
the number of training sentences was less than 40, VBEC greatly outperformed the MLBIC/MDL (A). With ML-BIC/MDL (A), an appropriate model structure was obtained by
BIC/M DL
maximizing an objective function lm
w.r.t. m based on BIC/MDL defined as:
# (? ? )
log T?(0) ,
(9)
2
where, l(O, m) denotes the likelihood of training data O for a model structure m, # (? ? )
denotes the number of free parameters for a set of states ?, and T?(0) denotes the total
frame number of training data for a set of states ?(0) in a root node, as shown in Figure 2.
The term # (?2 ? ) log T?(0) in Eq.(9) is regarded as a penalty term added to a likelihood, and
is dependent on the number of free parameters # (? ? ) and total frame number T?(0) of the
training data. ML-BIC/MDL (A) was based on the original definitions of BIC/MDL and
has been widely used in speech recognition [2] [5]. With such small amounts of training
data, there was a great difference between the total number of shared states with VBEC and
BIC/M DL
lm
= l(O, m) ?
Table 3: Training and recognition data
Training
Recognition
Continuous speech sentences (Acoustical Society of Japan)
100 city names (Japan Electronic Industry Development Association)
?????
??
??
??
????
??????????????
??????????????
??
?
?
??
???
????
??????????????
?????
Figure 4: Recognition rates according to the
amounts of training data based on the VBEC
and ML-BIC/MDL (A) and (B). The horizontal axis is scaled logarithmically.
???????????
????????????????????
???
????
???
????
??????????????
??????????????
??
?
??
???
????
??????????????
?????
Figure 5: Number of shared states according to the amounts of training data based on
the VBEC and ML-BIC/MDL (A) and (B).
The horizontal and vertical axes are scaled
logarithmically.
ML-BIC/MDL (A) (Figure 5). This suggests that VBEC, which does not use an asymptotic
assumption, determines the model structure more appropriately than the ML-BIC/MDL
(A), when the training data size is small.
Next, we adjusted the penalty term of ML-BIC/MDL in Eq. (9) so that the total numbers of
states for small amounts of data were as close as possible to those of VBEC (ML-BIC/MDL
(B) in Figure 5). Nevertheless, the recognition rates obtained by VBEC were about 15 %
better than those of ML-BIC/MDL (B) with fewer than 15 training sentences (Figure 4).
With such very small amounts of data, the VBEC and ML-BIC/MDL (B) model structures
were almost same (Figure 5). It is assumed that the effects of the posterior estimation and
the Bayesian prediction classification (Eq. (8)) suppressed the over-fitting of the models to
very small amounts of training data compared with the ML estimation and recognition in
ML-BIC/MDL (B).
With more than 100 training sentences, the recognition rates obtained by VBEC converged
asymptotically to those obtained by ML-BIC/MDL methods as the amounts of training data
became large.
In summary, VBEC performed as well or better for every amount of training data. This
advantage was due to the superior properties of VBEC, e.g., the appropriate determination
of the number of states and the suppression effect on over-fitting.
6.2
Influence of hyperparameter values on the quality of SST-HMMs
Throughout the construction of the model structure, the estimation of the posterior distribution, and recognition, we used a fixed combination of hyperparameter values, ? 0 = ? 0 =
0.01. In the small-scale experiments conducted in previous research [1] [3] [6] [8], the
selection of such values was not a major concern. However, when the scale of the target
application is large, the selection of hyperparameter values might affect the quality of the
models. Namely, the best or better values might differ greatly according to the amounts of
training data. Moreover, estimating appropriate hyperparameters with training SST-HMMs
needs so much time that it is impractical in speech recognition. Therefore, we examined
how robustly the SST-HMMs produced by VBEC performed against changes in the hyperparameter values with varying amounts of training data.
We varied the values of hyperparameters ? 0 and ? 0 from 0.0001 to 1, and examined the
speech recognition rates in two typical cases; one in which the amount of data was very
small (10 sentences) and one in which the amount was fairly large (150 sentences). Tables
Table 4: Recognition rates in each prior
distribution parameter when using training data of 10 sentences.
Table 5: Recognition rates in each prior
distribution parameter when using training data of 150 sentences.
?0
?0
100
10?1
10?2
10?3
10?4
100
1.0
2.2
31.2
60.3
66.5
10?1
66.3
65.9
66.1
66.2
66.6
?0
10?2
65.9
66.2
66.5
66.7
66.3
10?3
66.5
66.7
66.3
66.1
65.5
10?4
66.1
66.1
65.5
65.5
64.6
100
10?1
10?2
10?3
10?4
100
22.0
49.3
83.5
92.5
94.1
10?1
93.5
94.3
94.4
93.8
93.2
?0
10?2
94.0
93.9
93.2
93.3
92.3
10?3
93.1
93.3
92.3
92.5
92.3
10?4
92.3
92.5
92.3
92.4
92.2
4 and 5 show the recognition rates for each combination of hyperparameters. We can
see that the hyperparameter values for acceptable performance are broadly distributed for
both very small and fairly large amounts of training data. Moreover, roughly the ten best
recognition rates are highlighted in the tables. The combinations of hyperparameter values
that achieved the highlighted recognition rates were similar for the two different amounts of
training data. Namely, appropriate combinations of hyperparameter values can consistently
provide good performance levels regardless of the varying amounts of training data.
In summary, the hyperparameter values do not greatly influence the quality of the SSTHMMs. This suggests that it is not necessary to select the hyperparameter values very
carefully.
7 Conclusion
In this paper, we proposed VBEC, which constructs SST-HMMs based on the VB approach,
and recognizes speech based on the Bayesian prediction classification. With VBEC, the
model structure of SST-HMMs is adaptively determined according to the amounts of given
training data, and therefore a robust speech recognition system can be constructed. The first
experimental results, obtained by using real speech recognition tasks, showed the effectiveness of VBEC. In particular, when the training data size was small, VBEC significantly
outperformed conventional methods. The second experimental results suggested that it is
not necessary to select the hyperparameter values very carefully. From these results, we
conclude that VBEC provides a completely Bayesian framework for speech recognition
which effectively hundles the sparse data problem.
References
[1] H. Attias, ?A Variational Bayesian Framework for Graphical Models,? NIPS12, MIT Press,
(2000).
[2] W. Chou and W. Reichl, ?Decision Tree State Tying Based on Penalized Bayesian Information
Criterion,? Proc. ICASSP?99, vol. 1, pp. 345-348, (1999).
[3] Z. Ghahramani and M. J. Beal, ?Variational Inference for Bayesian Mixtures of Factor Analyzers,? NIPS12, MIT Press, (2000).
[4] J. J. Odell, ?The Use of Context in Large Vocabulary Speech Recognition,? PhD thesis, Cambridge University, (1995).
[5] K. Shinoda and T. Watanabe, ?Acoustic Modeling Based on the MDL Principle for Speech
Recognition,? Proc. EuroSpeech?97, vol. 1, pp. 99-102, (1997).
[6] N. Ueda and Z. Ghahramani, ?Bayesian Model Search for Mixture Models Based on Optimizing
Variational Bounds,? Neural Networks, vol. 15, pp. 1223-1241, (2002).
[7] S. Watanabe et. al., ?Constructing Shared-State Hidden Markov Models Based on a Bayesian
Approach,? Proc. ICSLP?02, vol. 4, pp. 2669-2672, (2002).
[8] S. Waterhouse et. al., ?Bayesian Methods for Mixture of Experts,? NIPS8, MIT Press, (1995).
| 2174 |@word briefly:1 retraining:2 covariance:3 score:2 selecting:2 hereafter:1 od:2 must:1 partition:1 enables:1 designed:1 update:1 leaf:2 fewer:1 provides:1 node:10 constructed:2 consists:2 fitting:2 roughly:1 seika:1 inappropriate:1 window:1 estimating:3 moreover:2 maximizes:1 tying:1 kind:1 spoken:2 st0:2 corporation:1 impractical:1 every:2 xd:1 scaled:2 partitioning:1 unit:5 bst:3 insufficiency:1 yd:1 might:2 examined:3 suggests:2 co:1 hmms:19 bjk:2 practical:2 procedure:1 significantly:1 word:1 close:1 selection:7 context:2 applying:1 ast:1 influence:2 optimize:1 conventional:4 a33:1 deterministic:1 yt:2 maximizing:3 center:1 go:1 regardless:1 independently:1 splitting:5 regarded:1 construction:1 target:1 logarithmically:2 recognition:38 jk:41 solved:2 calculate:1 decrease:1 removed:2 mentioned:2 complexity:1 predictive:3 completely:1 icassp:1 joint:1 k0:1 represented:2 quite:1 heuristic:3 widely:3 solve:1 otherwise:1 highlighted:2 ip:1 beal:1 advantage:2 sequence:5 propose:1 description:1 cluster:1 a11:1 ij:17 a22:1 phonemic:2 eq:4 nips12:2 qd:1 differ:1 merged:1 a12:1 icslp:1 generalization:1 clustered:2 minami:2 adjusted:1 hold:2 practically:1 great:1 viterbi:4 lm:2 major:1 estimation:6 proc:3 outperformed:2 applicable:1 prepare:1 successfully:1 city:2 mit:3 gaussian:13 varying:3 ax:3 odt:1 consistently:1 likelihood:4 mainly:1 greatly:3 digamma:1 suppression:1 chou:1 inference:1 dependent:2 nn:5 hidden:4 interested:1 arg:2 classification:6 among:1 development:1 initialize:1 fairly:2 construct:5 once:1 sampling:1 identical:1 represents:1 inflates:1 randomly:1 composed:1 gamma:1 phase:1 vowel:1 attempt:1 mdl:25 male:2 mixture:12 held:2 naonori:1 integral:1 necessary:3 tree:11 isolated:1 industry:1 modeling:5 assignment:3 subset:3 recognizing:1 conducted:2 too:1 eurospeech:1 answer:1 cho:1 adaptively:2 st:16 density:1 probabilistic:2 thesis:1 successively:2 choose:1 expert:1 japan:3 student:1 summarized:1 performed:3 try:1 root:4 closed:1 start:1 hf:1 contribution:1 became:1 phoneme:11 variance:1 correspond:1 yes:7 bayesian:25 produced:2 mfcc:2 classified:1 converged:1 explain:1 definition:1 failure:1 against:2 pp:4 hamming:1 stop:2 gain:1 carefully:2 horizontal:2 quality:3 name:2 effect:2 consisted:2 true:2 analytically:1 laboratory:1 iteratively:1 deal:1 maintained:1 criterion:6 m:1 tt:2 atsushi:1 variational:18 superior:1 attached:1 jp:1 khz:1 association:1 refer:1 cambridge:1 rd:1 analyzer:1 language:1 jj0:1 triphones:2 posterior:18 showed:1 optimizing:1 phonetic:8 binary:2 continue:1 vt:11 minimum:1 preceding:2 employed:3 r0:3 triphone:17 determine:1 maximize:1 monotonically:1 odell:1 multiple:1 kyoto:1 ntt:3 determination:2 calculation:1 prediction:6 expectation:1 iteration:1 represent:1 achieved:1 crucial:1 appropriately:1 ot:3 unlike:2 effectiveness:3 split:2 xj:1 bic:25 affect:1 fm:7 restrict:1 reduce:1 attias:1 shift:1 t0:2 penalty:2 soraku:1 speech:22 jj:3 ignored:1 sst:14 nasal:1 amount:24 prepared:2 ten:1 category:1 estimated:2 per:1 broadly:1 discrete:1 hyperparameter:11 vol:4 four:1 nevertheless:1 backward:2 asymptotically:1 almost:1 throughout:1 ueda:3 electronic:1 decision:6 acceptable:1 vb:14 bit:1 bound:1 datum:1 relatively:1 cslab:1 according:6 combination:4 conjugate:1 belonging:1 em:1 suppressed:1 count:1 available:1 apply:2 appropriate:8 spectral:2 robustly:1 robustness:1 original:1 denotes:13 clustering:7 dirichlet:1 recognizes:3 graphical:1 maintaining:1 ghahramani:2 especially:1 hikaridai:1 society:1 objective:4 question:6 added:1 diagonal:1 hmm:12 gun:1 acoustical:1 considers:1 rjk:2 length:1 index:1 insufficient:2 balance:2 unknown:1 vertical:1 markov:3 t:1 communication:1 precise:1 frame:6 varied:1 introduced:1 namely:3 sentence:10 acoustic:9 suggested:1 usually:2 below:1 including:1 nakamura:1 normality:1 axis:1 review:1 prior:8 waterhouse:1 asymptotic:4 shinji:1 occupation:2 versus:1 sufficient:1 kecl:1 principle:4 thresholding:1 summary:2 penalized:1 free:2 aij:4 sification:1 sparse:1 distributed:1 regard:1 calculated:2 vocabulary:1 valid:1 transition:4 forward:2 approximate:2 dealing:1 ml:19 assumed:2 conclude:1 continuous:1 latent:5 iterative:2 search:1 table:9 terminate:1 robust:1 obtaining:1 complex:1 japanese:2 constructing:2 hyperparameters:6 retrain:1 ny:5 watanabe:4 xl:1 answering:1 xt:10 concern:1 dl:2 cjk:4 quantization:1 sequential:1 effectively:2 phd:1 ch:1 determines:1 extracted:1 invalid:1 shared:7 change:2 experimentally:1 cst:1 determined:4 specifically:1 typical:1 preset:1 total:9 experimental:3 yasuhiro:1 select:2 evaluate:2 |
1,290 | 2,175 | The RA Scanner: Prediction of Rheumatoid
Joint Inflammation Based on Laser Imaging
1
Anton Schwaighofer1 2
TU Graz, Institute for Theoretical Computer Science
Inffeldgasse 16b, 8010 Graz, Austria
http://www.igi.tugraz.at/aschwaig
Volker Tresp, Peter Mayer
Siemens Corporate Technology, Department of Neural Computation
Otto-Hahn-Ring 6, 81739 Munich, Germany
http://www.tresp.org,[email protected]
2
Alexander K. Scheel, Gerhard Muller
?
University G?ottingen, Department of Medicine, Nephrology and Rheumatology
Robert-Koch-Stra?e 40, 37075 G?ottingen, Germany
[email protected],[email protected]
Abstract
We describe the RA scanner, a novel system for the examination of patients suffering from rheumatoid arthritis. The RA scanner is based on
a novel laser-based imaging technique which is sensitive to the optical
characteristics of finger joint tissue. Based on the laser images, finger
joints are classified according to whether the inflammatory status has
improved or worsened. To perform the classification task, various linear and kernel-based systems were implemented and their performances
were compared. Special emphasis was put on measures to reliably perform parameter tuning and evaluation, since only a very small data set
was available. Based on the results presented in this paper, it was concluded that the RA scanner permits a reliable classification of pathological finger joints, thus paving the way for a further development from
prototype to product stage.
1 Introduction
Rheumatoid arthritis (RA) is the most common inflammatory arthropathy with 1?2% of the
population being affected. This chronic, mostly progressive disease often leads to early disability and joint deformities. Recent studies have convincingly shown that early treatment
and therefore an early diagnosis is mandatory to prevent or at least delay joint destruction [2]. Unfortunately, long-term medication with disease modifying anti-rheumatic drugs
(DMARDs) often acts very slowly on clinical parameters of inflammation, making it difficult to find the right drug for a patient within adequate time. Conventional radiology,
such as magnetic resonance imaging (MRI) and ultrasound, may provide information on
soft tissue changes, yet these techniques are time-consuming and?in the case of MRI?
costly. New imaging techniques for RA diagnosis should thus be non-invasive, of low cost,
examiner independent and easy to use.
Following recent experiments on absorption and scattering coefficients of laser light in
joint tissue [6], a prototype laser imaging technique was developed [7]. As part of the prototype development, it became necessary to analyze if the rheumatic status of a finger joint
can be reliably classified on the basis of the laser images. Aim of this article is to provide an overview of this analysis. Employing different linear and kernel-based classifiers,
we will investigate the performance of the laser imaging technique to predict the status
of the rheumatic joint inflammation. Provided that the accuracy of the overall system is
sufficiently high, the imaging technique and the automatic inflammation classification can
be combined into a novel device that allows an inexpensive and objective assessment of
inflammatory joint changes.
The paper is organized as follows. In Sec. 2 we describe the RA scanner in more detail, as
well as the process of data acquisition. In Sec. 3 we describe the linear and kernel-based
classifiers used in the experiments. In Sec. 4 we describe how the methods were evaluated
and compared. We present experimental results in Sec. 5. Conclusions and an outlook are
given in Sec. 6.
2 The RA Scanner
The rheumatoid arthritis (RA) scanner provides a new medical imaging technique, developed specifically for the diagnosis of RA in finger joints. The RA scanner [7] allows the
in vivo trans-illumination of finger joints with laser light in the near infrared wavelength
range. The scattered light distribution is detected by a camera and is used to assess the
inflammatory status of the finger joint. Example images, taken from an inflamed joint and
from a healthy control, are shown in Fig. 1.
Starting out from the laser images, image pre-processing is used to obtain a description of
each laser image by nine numerical features. A brief description of the features is given in
Fig. 1. Furthermore for each finger joint examined, the circumference is measured using a
conventional measuring tape. The nine image features plus the joint circumference make up
the data that is used in the classification step of the RA scanner to predict the inflammatory
status of the joint.
2.1 Data Acquisition
One of the clinically important questions is to know as early as possible if a prescribed
medication improves the state of rheumatoid arthritis. Therefore the goal of the classification step in the RA scanner is to decide?based on features extracted from the laser
images?if there was an improvement of arthritis activity or if the joint inflammation remained unchanged or worsened.
The data for the development of the RA scanner stems from a study on 22 patients with
rheumatoid arthritis. Data from 72 finger joints were used for the study. All of these 72
finger joints were examined at baseline and during a follow-up visit after a mean duration of
42 days. Earlier data from an additional 20 patients had to be discarded since experimental
conditions were not controlled properly.
Each joint was examined and the clinical arthritis activity was classified from 0 (inactive,
not swollen, tender or warm) to 3 (very active) by a rheumatologist. The characteristics of
joint tissue was recorded by the above described laser imaging technique. In a preprocess-
(a) Laser image of a healthy finger joint
(b) Laser image of an inflamed finger
joint. The inflammation changes the joint
tissue?s absorption coefficient, giving a
darker image.
Figure 1: Two examples of the light distribution captured by the RA scanner. A laser beam
is sent through the finger joint (the finger tip is to the right, the palm is on the left), the light
distribution below the joint is captured by a CCD element. To calculate the features, first
a horizontal line near the vertical center of the finger joint is selected. The distribution of
light intensity along that line is bell-shaped. The features used in the classification task are
the maximum light intensity, the curvature of the light intensity at the maximum and seven
additional features based on higher moments of the intensity curve.
ing step nine features were derived from the distribution of the scattered laser light (see
Fig. 1). The tenth feature is the circumference of the finger joint.
Since there are high inter-individual variations in optical joint characteristics, it is not possible to tell the inflammatory status of a joint from one single image. Instead, special
emphasis was put on the intra-individual comparison of baseline and follow-up data. For
every joint examined, data from baseline and follow-up visit were compared and changes
in arthritis activity were rated as improvement, unchanged or worsening.
This rating divided the data into two classes: Class 1 contains the joints where an improvement of arthritis activity was observed (a total of 46 joints), and class 1 are the
joints that remained unchanged or worsened (a total of 26 joints). For all joints, the differences in feature values between baseline and follow-up visit were computed.
3 Classification Methods
In this section, we describe the employed linear and kernel-based classification methods,
where we focus on design issues.
3.1 Gaussian Process Classification (GPC)
In Gaussian processes, a function
f x
M
? w j K x x j ?
(1)
j 1
is described as a superposition of M kernel functions K x x j ? , defined for each of the
M training data points x j , with weight w j . The kernel functions are parameterized by the
vector ? ?0
?d . In two-class Gaussian process classification, the logistic transfer
function ? f x
1 e f x 1 is applied to the prediction of a Gaussian process to
produce an output which can be interpreted as ? x , the probability of the input x belonging
to class 1 [10].
In the experiment we chose the Gaussian kernel function
K x x j ?
?0 exp
1
x x j T diag ?21
?2d
2
1 x
x j
(2)
with input length scales ? 1 ?d where d is the dimension of the input space. diag ? 21
?2d
denotes a diagonal matrix with entries ? 21
?2d . For training the Gaussian process classifier
(that is, determining the posterior probabilities of the parameters w 1
wM ?0 ?d ) we
used a full Bayesian approach, implemented with Readford Neal?s freely available FBM
software.1
3.2 Gaussian Process Regression (GPR)
In
GPR we treat the classification problem as a regression problem with target values
1 1 , i.e. we do not apply the logistic transfer function as in the last subsection.
Any GP output 0 is treated as indicating an example from class 0, any output 0 as
an indicator for class 1.The disadvantage is that the GPR prediction cannot be treated as
a posterior class probability; the advantage is that the fast and non-iterative training algorithms for GPR can be applied. GPR for classification problems can be considered as
special cases of Fisher discriminant analysis with kernels [4] and of least squares support
vector machines [9].
The parameters ? ? 0
?d of the covariance function Eq. (2) were chosen by maximizing the posterior probability of ?, P ? t X ? P t X ? P ? , via a scaled conjugate
gradient method. Later on, this method will be referred to as ?GPR Bayesian?. Results are
also given for a simplified covariance function with ? 0 1, ?1 ?2
?d r, where
the common length scale r was chosen by cross-validation (later on referred to as ?GPR
crossval?).
3.3 Support Vector Machine (SVM)
The SVM is a maximum margin linear classifier. As in Sec. 3.2, the SVM classifies a
pattern according to the sign of f x in Eq. (1). The difference is that the weights w
w1
wM T in the SVM minimize the particular cost function [8]
wT Kw
M
? Ci 1
yi f xi
(3)
i 1
where sets all negative arguments to zero. Here, y i
1 1 is the class label for
training point x i . Ci 0 is a constant that determines the weight of errors on the training
data, and K is an M
M matrix containing the amplitudes of the kernel functions at the
training data, i.e. K i j K xi x j ? . The motivation for this cost function stems from statistical learning theory [8]. Many authors have previously obtained excellent classification
results by using the SVM. One particular feature of the SVM is the sparsity of the solution
vector w, that is, many elements w i are zero.
In the experiments, we used both an SVM with linear kernel (?SVM linear?) and an SVM
with a Gaussian kernel (?SVM Gaussian?), equivalent to the Gaussian process kernel
Eq. (2), with ? 0 1, ?1 ?2
?d r. The kernel parameter r was chosen by
cross-validation.
1 As a prior distribution for kernel parameter ?
0
we chose a Gamma distribution. ?1 ?d are samples of a hierarchical Gamma distribution. In FBM syntax, the prior is 0.05:0.5 x0.2:0.5:1.
Sampling from the posterior distribution was done by persistent hybrid Monte Carlo, following the
example of a 3-class problem in Neal [5].
To compensate for the unbalanced distribution of classes, the penalty term C i was chosen
to be 0 8 for the examples from the larger class and 1 for the smaller class. This was found
empirically to give the best balance of sensitivity and specificity (cf. Sec. 4). A formal
treatment of this issue can be found in Lin et al. [3].
3.4 Generalized Linear Model (GLM)
A GLM for binary responses is built up from a linear model for the input data, and the
model output f x w T x is in turn input to the link function. For Bernoulli distributions,
the natural link function [1] is the logistic transfer function ? f x 1 e f x
1 . The
overall output of the GLM ? f x computes ? x , the probability of the input x belonging
to class 1. Training of the linear model was done by iteratively re-weighted least squares
(IRLS).
4 Training and Evaluation
One of the challenges in developing the classification system for the RA scanner is the low
number of training examples available. Data was collected through an extensive medical
study, but only data from 72 fingers were found to be suitable for further use. Further
data can only be acquired in carefully controlled future studies, once the initial prototype
method has proven sufficiently successful.
Training From the currently available 72 training examples, classifiers need to be trained
and evaluated reliably. Part of the standard methodology for small data sets is N-fold crossvalidation, where the data are partitioned into N equally sized sets and the system is trained
on N 1 of those sets and tested on the Nth data set left out. Since we wish to make use of
as much training data as possible, N 36 seemed the appropriate choice 2 , giving test sets
with two examples in each iteration. For some of the methods model parameter needed
to be tuned (for example, choosing SVM kernel width), where again cross-validation is
employed. The nested cross-validation ensures that in no case any of the test examples is
used for training or to tune parameters, leading to the following procedure:
Run 36 fold CV
For Bayesian methods or methods without tunable parameters
(SVM linear, GPC, GPR Bayesian, GLM):
Use full training set to tune and train classifier
For Non-Bayesian methods (SVM Gaussian, GPR crossval):
Run 35 fold CV on the training set
choose parameters to minimise CV error
train classifier with chosen parameters
evaluate the classifier on the 2 example test set
Significance Tests In order to compare the performance of two given classification methods, one usually employs statistical hypothesis testing. We use here a test that is best suited
for small test sets, since it takes into account the outcome on the test examples one by one,
thus matching our above described 36-fold cross validation scheme perfectly. A similar
test has been used by Yang and Liu [11] to compare text categorization methods.
Basis of the test are two counts b (how many examples in the test set were correctly classified by method B, but misclassified by method A) and c (number of examples misclassified
by B, correctly classified by A). We assume that examples misclassified (resp. correctly
classified) by both A and B do not contribute to the performance difference. We take the
2 Thus,
it is equivalent to a leave-one-out scheme, yet with only half the time consumption.
Method
Error rate
GLM
GLM, reduced feature set
GPR Bayesian
GPR crossval
GPC
SVM linear
SVM linear, reduced feature set
SVM Gaussian
20 83%
16 67%
13 89%
22 22%
23 61%
22 22%
16 67%
20 83%
Table 1: Error rates of different classification methods on the rheumatoid arthritis prediction problem. All error rates have been computed by 36-fold cross-validation. ?Reduced
feature set? indicates experiments where a priori feature selection has been done
counts b and c as the sufficient statistics of a binomial random variable with parameter ?,
where ? is the proportion of cases where method A performs better than method B.
The null hypothesis H0 is that the parameter ? 0 5, that is, both methods A and B have
the same performance. Hypothesis H 1 is that ? 0 5. The test statistics under the null
hypothesis is the Binomial distribution Bi i b c ?) with parameter ? 0 5. We reject
the null hypothesis if the probability of observing a count k c under the null hypothesis
P k c ?bi cc Bi i b c ? 0 5 is sufficiently small.
ROC Curves In medical diagnosis, biometrics and other areas, the common means of
assessing a classification method is the receiver operating characteristics (ROC) curve. An
ROC curve plots sensitivity versus 1-specificity 3 for different thresholds of the classifier
output. Based on the ROC curve it can be decided how many false positives resp. false
negatives one is willing to tolerate, thus helping to tune the classifier threshold to best suit
a certain application.
Acquiring the ROC curve typically requires the classifier output on an independent test set.
We instead use the union of all test set outputs in the cross-validation routine. This means
that the ROC curve is based on outputs of slightly different models, yet this still seems to
be the most suitable solution for such few data. For all classifiers we assess the area of the
ROC curve and the cross-validation error rate. Here the above mentioned threshold on the
classifier output is chosen such that sensitivity equals specificity.
5 Results
Tab. 1 lists error rates for all methods listed in Sec. 3. Gaussian process regression (GPR
Bayesian) with an error rate of 14% clearly outperforms all other methods, which all
achieve comparable error rates in the range of 20
24%. We attribute the good performance of GPR to its inherent feature relevance detection, which is done by adapting the
length scales ?i in the covariance function Eq. (2), i.e. a large ? i means that the i-th feature
is essentially ignored.
Surprisingly, Gaussian process classification implemented with Markov chain Monte Carlo
sampling (GPC) showed rather poor performance. We currently have no clear explanation
for this fact. We found no indications of convergence problems, furthermore we achieved
similar results with different sampling schemes.
In an additional experiment we wanted to find out if classification results could be improved
3 sensitivity
true positives
true positives false negatives
specificity
true negatives
true negatives false positives
1
0.9
0.8
Sensitivity
0.7
0.6
0.5
0.4
0.3
0.2
GPR Bayesian
GLM, reduced feature set
SVM linear, reduced feature set
0.1
0
0
0.2
0.4
0.6
1?Specificity
0.8
1
Figure 2: ROC curves of the best classification methods, both on the full data set and on
a reduced data set where a priori feature selection was used to retain only the three most
relevant features. Integrating the area under the ROC curves gives similar results for all
three methods, with an area of 0 86 for SVM linear and GLM, and 0 84 for GPR Bayesian
by using only a subset of input features 4 . We found that only the performance of the two
linear classifiers (GLM and SVM linear) could be improved by the input feature selection.
Both now achieve an error rate of 16 67%, which is slightly worse than GPR on the full
feature set (see Tab. 1).
Significance Tests Using the statistical hypothesis test described in the previous section,
we compared all classification methods pairwise. It turned out the three best methods
(GPR Bayesian, and GLM and SVM linear with reduced feature set) perform better than
all other methods at a confidence level of 90% or more. Amongst the three best methods,
no significant difference could be observed.
ROC Curves For the three best classification methods (GPR Bayesian, and GLM and
SVM linear with reduced feature set), we have plotted the receiver operating characteristics
(ROC) curve in Fig. 2. According to the ROC curve a sensitivity of 80% can be achieved
with a specificity at around 90%. GPR Bayesian seems to give best results, both in terms
of error rate and shape of the ROC curve.
Summary To summarize, when the full set of features was used, best performance was
obtained with GPR Bayesian. We attribute this to the inherent input relevance detection
mechanisms of this approach. Comparable yet slightly worse results could be achieved
by performing feature selection a priori and reducing the number of input features to the
three most significant ones. In particular, the error rates of linear classifiers (GLM and
linear SVM) improved by this feature selection, whereas more complex classifiers did not
benefit. We can draw the important conclusion that, using the best classifiers, a sensitivity
of 80% can be reached at a specificity of approximately 90%.
6 Conclusions
In this paper we have reported results of the analysis of a prototype medical imaging system, the RA scanner. Aim of the RA scanner is to detect soft tissue changes in finger joints,
4 This was done with the input relevance detection algorithm of the neural network tool SENN,
a variant of sequential backward elimination where the feature that least affects the neural network
output is removed. The feature set was reduced to the three most relevant ones.
which occur in early stages of rheumatoid arthritis (RA). Basis of the RA scanner is a novel
laser imaging technique that is sensitive to inflammatory soft tissue changes.
We have analyzed whether the laser images are suitable for an accurate prediction of the
inflammatory status of a finger joint, and which classification methods are best suited for
this task. Out of a set of linear and kernel-based classification methods, Gaussian processes
regression performed best, followed closely by generalized linear models and the linear
support vector machine, the latter two operating on a reduced feature set. In particular, we
have shown how parameter tuning and classifier training can be done on basis of the scarce
available data. For the RA prediction task, we achieved a sensitivity of 80% at a specificity
of approximately 90%. These results show that a further development of the RA scanner is
desirable.
In the present study the inflammatory status is assessed by a rheumatologist, taking into
account the patients subjective degree of pain. Thus we may expect a certain degree of label
noise in the data we have trained the classification system on. Further developments of the
classification system in the RA scanner will thus incorporate information from established
medical imaging systems such as magnetic resonance imaging (MRI). MRI is known to
provide accurate information about soft tissue changes in finger joints, yet is too costly to
be routinely used for RA diagnosis. By incorporating MRI results into the RA scanner?s
classification system, we expect to significantly improve the overall accuracy.
Acknowledgments AS gratefully acknowledges support through an Ernst-von-Siemens
scholarship. Thanks go to Radford Neal for making his FBM software available to the
public, and to Ian Nabney and Chris Bishop for the Netlab toolbox.
References
[1] Fahrmeir, L. and Tutz, G. Multivariate Statistical Modelling Based on Generalized Linear
Models. Springer Verlag, 2nd edn., 2001.
[2] Kim, J. and Weisman, M. When does rheumatoid arthritis begin and why do we need to know?
Arthritis and Rheumatism, 43:473?482, 2000.
[3] Lin, Y., Lee, Y., and Wahba, G. Support vector machines for classification in nonstandard
situations. Tech. Rep. 1016, Department of Statistics, University of Wisconsin, Madison, WI,
USA, 2000.
[4] Mika, S., R?atsch, G., Weston, J., Sch?olkopf, B., Smola, A. J., and M?uller, K.-R. Invariant
feature extraction and classification in kernel spaces. In S. A. Solla, T. K. Leen, and K.-R.
M?uller, eds., Advances in Neural Information Processing Systems 12. MIT Press, 2000.
[5] Neal, R. M. Monte carlo implementation of gaussian process models for bayesian regression
and classification. Tech. Rep. 9702, Department of Statistics, University of Toronto, 1997.
[6] Prapavat, V., Runge, W., Krause, A., Beuthan, J., and M?uller, G. A. Bestimmung von gewebeoptischen Eigenschaften eines Gelenksystems im Fr?uhstadium der rheumatoiden Arthritis (in
vitro). Minimal Invasive Medizin, 8:7?16, 1997.
[7] Scheel, A. K., Krause, A., Mesecke-von Rheinbaben, I., Metzger, G., Rost, H., Tresp, V., Mayer,
P., Reuss-Borst, M., and M?uller, G. A. Assessment of proximal finger joint inflammation in
patients with rheumatoid arthritis, using a novel laser-based imaging technique. Arthritis and
Rheumatism, 46(5):1177?1184, 2002.
[8] Sch?olkopf, B. and Smola, A. J. Learning with Kernels. MIT Press, 2002.
[9] Van Gestel, T., Suykens, J. A., Lanckriet, G., Lambrechts, A., De Moor, B., and Vandewalle,
J. Bayesian framework for least-squares support vector machine classifiers, gaussian processes
and kernel fisher discriminant analysis. Neural Computation, 14(5):1115?1147, 2002.
[10] Williams, C. K. and Barber, D. Bayesian classification with gaussian processes. IEEE Transactions on Pattern Analysis and Machine Intelligence, 20(12):1342?1351, 1998.
[11] Yang, Y. and Liu, X. A re-examination of text categorization methods. In Proceedings of ACM
SIGIR 1999. ACM Press, 1999.
| 2175 |@word mri:5 proportion:1 seems:2 nd:1 willing:1 covariance:3 outlook:1 moment:1 initial:1 liu:2 contains:1 tuned:1 outperforms:1 subjective:1 worsening:1 yet:5 numerical:1 shape:1 wanted:1 plot:1 half:1 selected:1 device:1 intelligence:1 provides:1 contribute:1 toronto:1 org:1 along:1 crossval:3 persistent:1 acquired:1 pairwise:1 x0:1 inter:1 ra:25 borst:1 provided:1 classifies:1 begin:1 null:4 interpreted:1 developed:2 every:1 act:1 classifier:19 inflamed:2 scaled:1 control:1 medical:5 positive:4 treat:1 approximately:2 mika:1 plus:1 emphasis:2 chose:2 examined:4 range:2 bi:3 decided:1 acknowledgment:1 camera:1 testing:1 union:1 examiner:1 procedure:1 area:4 drug:2 bell:1 reject:1 adapting:1 matching:1 significantly:1 pre:1 integrating:1 aschwaig:1 specificity:8 confidence:1 cannot:1 selection:5 put:2 www:2 conventional:2 equivalent:2 chronic:1 circumference:3 center:1 maximizing:1 go:1 starting:1 duration:1 williams:1 sigir:1 his:1 population:1 variation:1 resp:2 target:1 gerhard:1 edn:1 hypothesis:7 rheumatoid:10 lanckriet:1 element:2 infrared:1 observed:2 calculate:1 graz:2 ensures:1 solla:1 removed:1 disease:2 mentioned:1 trained:3 basis:4 joint:41 worsened:3 various:1 finger:21 routinely:1 laser:19 train:2 fast:1 describe:5 monte:3 detected:1 tell:1 choosing:1 outcome:1 h0:1 larger:1 otto:1 statistic:4 radiology:1 gp:1 runge:1 advantage:1 indication:1 metzger:1 product:1 fr:1 tu:1 relevant:2 turned:1 stra:1 ernst:1 achieve:2 description:2 olkopf:2 crossvalidation:1 convergence:1 assessing:1 produce:1 categorization:2 ring:1 leave:1 measured:1 eq:4 implemented:3 closely:1 attribute:2 modifying:1 elimination:1 public:1 goettingen:1 absorption:2 im:1 helping:1 scanner:19 koch:1 sufficiently:3 considered:1 around:1 exp:1 predict:2 early:5 label:2 currently:2 superposition:1 healthy:2 sensitive:2 tool:1 weighted:1 moor:1 uller:4 mit:2 clearly:1 gaussian:18 aim:2 rather:1 volker:1 derived:1 focus:1 deformity:1 improvement:3 properly:1 bernoulli:1 indicates:1 modelling:1 tech:2 medication:2 baseline:4 detect:1 kim:1 destruction:1 typically:1 misclassified:3 germany:2 overall:3 classification:31 issue:2 priori:3 development:5 resonance:2 special:3 equal:1 once:1 shaped:1 extraction:1 sampling:3 progressive:1 kw:1 future:1 inherent:2 employ:1 few:1 pathological:1 gamma:2 individual:2 suit:1 detection:3 investigate:1 intra:1 evaluation:2 analyzed:1 light:9 chain:1 accurate:2 necessary:1 biometrics:1 re:2 plotted:1 theoretical:1 minimal:1 soft:4 earlier:1 inflammation:7 disadvantage:1 measuring:1 cost:3 entry:1 subset:1 delay:1 successful:1 vandewalle:1 too:1 reported:1 proximal:1 combined:1 thanks:1 sensitivity:8 retain:1 lee:1 tip:1 w1:1 again:1 von:3 recorded:1 containing:1 choose:1 slowly:1 worse:2 leading:1 account:2 de:4 sec:8 coefficient:2 igi:1 later:2 performed:1 analyze:1 observing:1 tab:2 wm:2 reached:1 vivo:1 ass:2 square:3 minimize:1 accuracy:2 became:1 gestel:1 characteristic:5 preprocess:1 anton:1 bayesian:16 irls:1 carlo:3 cc:1 tissue:8 classified:6 nonstandard:1 ed:1 inexpensive:1 acquisition:2 invasive:2 tunable:1 treatment:2 austria:1 subsection:1 improves:1 organized:1 amplitude:1 routine:1 carefully:1 scattering:1 higher:1 tolerate:1 day:1 follow:4 methodology:1 response:1 improved:4 leen:1 evaluated:2 done:6 furthermore:2 stage:2 smola:2 horizontal:1 assessment:2 logistic:3 usa:1 true:4 iteratively:1 neal:4 during:1 width:1 generalized:3 syntax:1 performs:1 image:13 novel:5 common:3 empirically:1 overview:1 vitro:1 significant:2 cv:3 tuning:2 automatic:1 ottingen:2 gratefully:1 had:1 operating:3 curvature:1 posterior:4 multivariate:1 recent:2 showed:1 mandatory:1 certain:2 verlag:1 binary:1 rep:2 yi:1 muller:1 der:1 captured:2 additional:3 employed:2 freely:1 inflammatory:9 full:5 corporate:1 desirable:1 stem:2 fbm:3 ing:1 clinical:2 long:1 cross:8 compensate:1 divided:1 lin:2 equally:1 visit:3 controlled:2 prediction:6 variant:1 regression:5 patient:6 essentially:1 iteration:1 kernel:19 arthritis:16 achieved:4 suykens:1 beam:1 whereas:1 krause:2 concluded:1 sch:2 med:1 sent:1 near:2 yang:2 easy:1 affect:1 perfectly:1 wahba:1 prototype:5 minimise:1 inactive:1 whether:2 penalty:1 peter:2 nine:3 tape:1 adequate:1 ignored:1 gpc:4 clear:1 listed:1 tune:3 reduced:10 http:2 senn:1 sign:1 correctly:3 diagnosis:5 affected:1 threshold:3 prevent:1 tenth:1 backward:1 imaging:14 run:2 parameterized:1 decide:1 draw:1 comparable:2 netlab:1 followed:1 nabney:1 fold:5 activity:4 occur:1 software:2 argument:1 prescribed:1 performing:1 optical:2 department:4 munich:1 according:3 palm:1 developing:1 clinically:1 poor:1 belonging:2 conjugate:1 smaller:1 slightly:3 partitioned:1 wi:1 making:2 invariant:1 glm:12 taken:1 previously:1 turn:1 count:3 mechanism:1 needed:1 know:2 available:6 permit:1 apply:1 hierarchical:1 appropriate:1 magnetic:2 rost:1 paving:1 denotes:1 binomial:2 cf:1 tugraz:1 ccd:1 madison:1 medicine:1 giving:2 scholarship:1 hahn:1 unchanged:3 objective:1 question:1 costly:2 diagonal:1 disability:1 gradient:1 amongst:1 pain:1 link:2 consumption:1 chris:1 seven:1 barber:1 collected:1 discriminant:2 length:3 balance:1 difficult:1 mostly:1 unfortunately:1 robert:1 negative:5 design:1 reliably:3 implementation:1 perform:3 vertical:1 markov:1 ultrasound:1 discarded:1 anti:1 situation:1 intensity:4 rating:1 toolbox:1 extensive:1 mayer:3 established:1 trans:1 below:1 pattern:2 usually:1 sparsity:1 challenge:1 convincingly:1 summarize:1 built:1 reliable:1 explanation:1 suitable:3 treated:2 examination:2 warm:1 hybrid:1 indicator:1 natural:1 scarce:1 nth:1 scheme:3 improve:1 technology:1 brief:1 rated:1 acknowledges:1 tresp:3 text:2 prior:2 tutz:1 determining:1 wisconsin:1 expect:2 proven:1 versus:1 validation:8 degree:2 sufficient:1 article:1 summary:1 surprisingly:1 last:1 formal:1 rheumatism:2 institute:1 taking:1 fahrmeir:1 mchp:1 benefit:1 van:1 curve:14 dimension:1 computes:1 seemed:1 author:1 simplified:1 employing:1 transaction:1 uni:1 status:8 active:1 gwdg:1 receiver:2 consuming:1 xi:2 iterative:1 why:1 table:1 transfer:3 excellent:1 complex:1 diag:2 did:1 significance:2 motivation:1 noise:1 suffering:1 fig:4 referred:2 roc:13 scattered:2 darker:1 wish:1 gpr:20 ian:1 remained:2 bishop:1 list:1 svm:22 incorporating:1 false:4 sequential:1 ci:2 illumination:1 margin:1 suited:2 wavelength:1 acquiring:1 radford:1 nested:1 springer:1 determines:1 extracted:1 acm:2 weston:1 goal:1 sized:1 tender:1 fisher:2 change:7 specifically:1 reducing:1 wt:1 total:2 experimental:2 siemens:3 atsch:1 indicating:1 support:6 latter:1 unbalanced:1 alexander:1 relevance:3 assessed:1 incorporate:1 evaluate:1 tested:1 |
1,291 | 2,176 | Automatic Derivation of Statistical Algorithms:
The EM Family and Beyond
Alexander G. Gray
Carnegie Mellon University
[email protected]
Bernd Fischer and Johann Schumann
RIACS / NASA Ames
fisch,schumann @email.arc.nasa.gov
Wray Buntine
Helsinki Institute for IT
[email protected]
Abstract
Machine learning has reached a point where many probabilistic methods can be understood as variations, extensions and combinations of a
much smaller set of abstract themes, e.g., as different instances of the
EM algorithm. This enables the systematic derivation of algorithms customized for different models. Here, we describe the AUTO BAYES system which takes a high-level statistical model specification, uses powerful symbolic techniques based on schema-based program synthesis and
computer algebra to derive an efficient specialized algorithm for learning
that model, and generates executable code implementing that algorithm.
This capability is far beyond that of code collections such as Matlab toolboxes or even tools for model-independent optimization such as BUGS
for Gibbs sampling: complex new algorithms can be generated without new programming, algorithms can be highly specialized and tightly
crafted for the exact structure of the model and data, and efficient and
commented code can be generated for different languages or systems.
We present automatically-derived algorithms ranging from closed-form
solutions of Bayesian textbook problems to recently-proposed EM algorithms for clustering, regression, and a multinomial form of PCA.
1 Automatic Derivation of Statistical Algorithms
Overview. We describe a symbolic program synthesis system which works as a ?statistical
algorithm compiler:? it compiles a statistical model specification into a custom algorithm
design and from that further down into a working program implementing the algorithm
design. This system, AUTO BAYES, can be loosely thought of as ?part theorem prover, part
Mathematica, part learning textbook, and part Numerical Recipes.? It provides much more
flexibility than a fixed code repository such as a Matlab toolbox, and allows the creation
of efficient algorithms which have never before been implemented, or even written down.
AUTO BAYES is intended to automate the more routine application of complex methods in
novel contexts. For example, recent multinomial extensions to PCA [2, 4] can be derived
in this way.
The algorithm design problem. Given a dataset and a task, creating a learning method
can be characterized by two main questions: 1. What is the model? 2. What algorithm will
optimize the model parameters? The statistical algorithm (i.e., a parameter optimization
algorithm for the statistical model) can then be implemented manually. The system in
this paper answers the algorithm question given that the user has chosen a model for the
data,and continues through to implementation. Performing this task at the state-of-the-art
level requires an intertwined meld of probability theory, computational mathematics, and
software engineering. However, a number of factors unite to allow us to solve the algorithm
design problem computationally: 1. The existence of fundamental building blocks (e.g.,
standardized probability distributions, standard optimization procedures, and generic data
structures). 2. The existence of common representations (i.e., graphical models [3, 13] and
program schemas). 3. The formalization of schema applicability constraints as guards. 1
The challenges of algorithm design. The design problem has an inherently combinatorial
nature, since subparts of a function may be optimized recursively and in different ways.
It also involves the use of new data structures or approximations to gain performance. As
the research in statistical algorithms advances, its creative focus should move beyond the
ultimately mechanical aspects and towards extending the abstract applicability of already
existing schemas (algorithmic principles like EM), improving schemas in ways that generalize across anything they can be applied to, and inventing radically new schemas.
2 Combining Schema-based Synthesis and Bayesian Networks
Statistical Models.
Externally,
A
UTO
B
AYES
has
the
look
and feel of
2 const int n_points as ?nr. of data points?
a compiler. Users specify their model
3
with 0 < n_points;
4 const int n_classes := 3 as ?nr. classes?
of interest in a high-level specification
5
with 0 < n_classes
language (as opposed to a program6
with n_classes << n_points;
ming language). The figure shows the
7 double phi(1..n_classes) as ?weights?
specification of the mixture of Gaus8
with 1 = sum(I := 1..n_classes, phi(I));
9 double mu(1..n_classes);
sians example used throughout this
9 double sigma(1..n_classes);
paper.2 Note the constraint that the
10 int c(1..n_points) as ?class labels?;
sum of the class probabilities must
11 c ? disc(vec(I := 1..n_classes, phi(I)));
equal one (line 8) along with others
12 data double x(1..n_points) as ?data?;
(lines 3 and 5) that make optimization
13 x(I) ? gauss(mu(c(I)), sigma(c(I)));
of the model well-defined. Also note
14 max pr(x| phi,mu,sigma ) wrt phi,mu,sigma ;
the ability to specify assumptions of
the kind in line 6, which may be used by some algorithms. The last line specifies the goal
with respect to the painference task: maximize the conditional probability pr
rameters , , and . Note that moving the parameters across to the left of the conditioning
bar converts this from a maximum likelihood to a maximum a posteriori problem.
1 model mog as ?Mixture of Gaussians?;
Computational logic and theorem proving. Internally, AUTO BAYES uses a class of techniques known as computational logic which has its roots in automated theorem proving.
AUTO BAYES begins with an initial goal and a set of initial assertions, or axioms, and adds
new assertions, or theorems, by repeated application of the axioms, until the goal is proven.
In our context, the goal is given by the input model; the derived algorithms are side effects
of constructive theorems proving the existence of algorithms for the goal.
1
Schema guards vary widely; for example, compare Nead-Melder simplex or simulated annealing (which require only function evaluation), conjugate gradient (which require both Jacobian and
Hessian), EM and its variational extension [6] (which require a latent-variable structure model).
2
Here, keywords have been underlined and line numbers have been added for reference in the text.
The as-keyword allows annotations to variables which end up in the generated code?s comments.
Also, n classes has been set to three (line 4), while n points is left unspecified. The class
variable and single data variable are vectors, which defines them as i.i.d.
Computer algebra. The first core element which makes automatic algorithm derivation
feasible is the fact that we can mechanize the required symbol manipulation, using computer algebra methods. General symbolic differentiation and expression simplification are
capabilities fundamental to our approach. AUTO BAYES contains a computer algebra engine using term rewrite rules which are an efficient mechanism for substitution of equal
quantities or expressions and thus well-suited for this task.3
Schema-based synthesis. The computational cost of full-blown theorem proving grinds
simple tasks to a halt while elementary and intermediate facts are reinvented from scratch.
To achieve the scale of deduction required by algorithm derivation, we thus follow a
schema-based synthesis technique which breaks away from strict theorem proving. Instead,
we formalize high-level domain knowledge, such as the general EM strategy, as schemas.
A schema combines a generic code fragment with explicitly specified preconditions which
describe the applicability of the code fragment. The second core element which makes
automatic algorithm derivation feasible is the fact that we can use Bayesian networks to
efficiently encode the preconditions of complex algorithms such as EM.
First-order logic representation of Bayesian netNclasses
works. A first-order logic representation of Bayesian
?
?
networks was developed by Haddawy [7]. In this
framework, random variables are represented by
functor symbols and indexes (i.e., specific instances
?
x
c
of i.i.d. vectors) are represented as functor arguments.
discrete
gauss
Nclasses
Since unknown index values can be represented by
Npoints
implicitly universally quantified Prolog variables, this
approach allows a compact encoding of networks involving i.i.d. variables or plates [3]; the
figure shows the initial network for our running example. Moreover, such networks correspond to backtrack-free datalog programs, allowing the dependencies to be efficiently
computed. We have extended the framework to work with non-ground probability queries
since we seek to determine probabilities over entire i.i.d. vectors and matrices. Tests for independence on these indexed Bayesian networks are easily developed in Lauritzen?s framework which uses ancestral sets and set separation [9] and is more amenable to a theorem
prover than the double negatives of the more widely known d-separation criteria. Given a
Bayesian network, some probabilities can easily be extracted by enumerating the component probabilities at each node:
Lemma
1. Let
be
sets
of variables over a Bayesian network with
. Then
and parents
hold 4 in the corresponding dependency
descendents
graph iff the following probability statement holds:
parents
!
"$#&%
&('
parents
(')+*
Symbolic probabilistic inference. How can probabilities not satisfying these conditions
be converted to symbolic expressions? While many general schemes for inference on networks exist, our principal hurdle is the need to perform this over symbolic expressions incorporating real and integer variables from disparate real or infinite-discrete distributions.
For instance, we might wish to compute the full maximum a posteriori probability for
the mean and variance vectors of a Gaussian mixture model under a Bayesian framework.
While the sum-product framework of [8] is perhaps closer to our formulation, we have out
of necessity developed another scheme that lets us extract probabilities on a large class of
mixed discrete and real, potentially indexed variables, where no integrals are needed and
3
Popular symbolic packages such as Mathematica contain known errors allowing unsound derivations; they also lack the support for reasoning with vector and matrix quantities.
-,
.
/,
.
4
Note that
descendents
and
parents
.
all marginalization is done by summing out discrete variables. We give the non-indexed
case below; this is readily extended to indexed variables (i.e., vectors).
Lemma
2.
descendents
holds and ancestors
is independent of given
iff
there
exists
a
set
of
variables
that Lemma 1 holds if we replace
by
. Moreover, the unique minimal such
set
satisfying
these
conditions
is
given
by
ancestors
ancestors
such that ancestors
is independent
Lemma
be a subset of
descendents
3. Let
of
&
ancestors
given
.
Then
Lemma
2
holds
if
we
replace
by
and
by . Moreover, there is a unique maximal set
satisfying these
conditions.
Lemma 2 lets us evaluate a probability by a summation:
&
' +
& )
" (# Dom %
while Lemma 3 lets us evaluate a probability by a summation and a ratio:
&
Since the lemmas also show minimality of the sets and
&
&
, they also give the minimal
conditions under which a probability can be evaluated by discrete summation without integration. These inference lemmas are operationalized as network decomposition schemas.
However, we usually attempt to decompose a probability into independent components
before applying this schema.
3 The AUTOBAYES System ? Implementation Outline
Levels of representation. Internally, our system uses three conceptually different levels of
representation. Probabilities (including logarithmic and conditional probabilities) are the
most abstract level. They are processed via methods for Bayesian network decomposition
or match with core algorithms such as EM. Formulae are introduced when probabilities of
the form parents
are detected, either in the initial network, or after the application of network decompositions. Atomic probabilities (i.e., is a single variable) are
directly replaced by formulae based on the given distribution and its parameters. General
probabilities are decomposed into sums and products of the respective atomic probabilities. Formulae are ready for immediate optimization using symbolic or numeric methods
but sometimes they can be decomposed further into independent subproblems. Finally, we
use imperative intermediate code as the lowest level to represent both program fragments
within the schemas as well as the completely constructed programs. All transformations
we apply operate on or between these levels.
Transformations for optimization. A number of different kinds of transformations are
available. Decomposition of a problem into independent subproblems is always done. Decomposition of probabilities is driven by the Bayesian network; we have a separate system
for handling decomposition of formulae. A formula can be decomposed along a loop, e.g.,
for "!
? is transformed into a for-loop over subproblems
the problem
?optimize
for !
$#&%
? is transformed
?optimize for !
.? More commonly,
?optimize
into the two subprograms ?optimize for !
? and ?optimize for %
.? The lemmas
given earlier are applied to change the level of representation and are thus used for simplification of probabilities. Examples of general expression simplification include simplifying
the log of a formula, moving a summation inwards, and so on. When necessary, symbolic
differentiation is performed. In the initial specification or in intermediate representations,
likelihoods (i.e., subexpressions of the form %
) are identified and simplified into linear expression with terms such as mean
and mean
. The statistical
algorithm schemas currently implemented include EM, k-means, and discrete model selection. Adding a Gibbs sampling schema would yield functionality comparable to that of
BUGS [14]. Usually, the schemas require a particular form of the probabilities involved;
they are thus tightly coupled to the decomposition and simplification transformations. For
is
example, EM is a way of dealing with situation where Lemma 2 applies but where
indexed identically to the data.
Code and test generation. From the intermediate code, code in a particular target language may be generated. Currently, AUTO BAYES can generate C++ and C which can be
used in a stand-alone fashion or linked into Octave or Matlab (as a mex file). During this
code-generation phase, most of the vector and matrix expressions are converted into forloops, and various code optimizations are performed which are impossible for a standard
compiler. Our tool does not only generate efficient code, but also highly readable, documented programs: model- and algorithm-specific comments are generated automatically
during the synthesis phase. For most examples, roughly 30% of the produced lines are
comments. These comments provide explanation of the algorithm?s derivation. A generated HTML software design document with navigation capabilities facilitates code understanding and reading. AUTO BAYES also automatically generates a program for sampling
from the specified model, so that closed-loop testing with synthetic data of the assumed
distributions can be done. This can be done using simple forward sampling.
4 Example: Deriving the EM Algorithm for Gaussian Mixtures
1. User specifies model. First, the user specifies the model as shown in Section 2.
2. System parses model to obtain underlying Bayes net. From the model, the underlying
Bayesian network is derived and represented internally as a directed graph. For visualization, AUTO BAYES can also produce a graph drawing as shown in Section 2.
3. System observes hidden-variable structure in Bayesian network. The system attempts to decompose the optimization goal into independent parts, but finds that it cannot.
However, it then finds that the probability in the initial optimization statement matches the
conditions of Lemma 2 and that the network describes a latent variable model.
4. System invokes abstract EM
schema max Pr
wrt
family schema. This triggers the
...
EM-schema, whose overall structure
C = ?[initialize
];)
is shown. The syntactic structure of
while
+ wrt ;
the current subproblem must match
/* M-step */ max Pr
the first argument of the schema;
/* E-step */ calculate Pr
;
if additional applicability constraints
?
(not shown here) hold, this schema is
executed. It constructs a piece of code which is returned in the variable . This code fragment can contain recursive calls to other schemas (denoted by !
" ) which return code for
subproblems which then is inserted into the schema, such as converging, a generic con
vergence criterion here imposed over the variables . Note that the schema actually
implements an ME-algorithm (i.e., starts the loop with the M-step) because the initialization already serves as an E-step.
The system identifies the discrete variable # as the
#
. For representation of the distribution of the hidden
single hidden variable, i.e., $
variable a matrix % is generated, where % '& is the probability that the ( -th point falls into
the ) -th class. AUTO BAYES then constructs the new distribution c(I) ? disc(vec(J
:= 1..n classes, q(I, J)) which replaces the original distribution in the following
recursive calls of AUTO BAYES.
while converging
for
for &
Pr &
"
!
!
max Pr
wrt #
5. E-step: System performs marginalization. The freshly introduced distribution for
#
implies that # can be eliminated from
the objective function by summing over % $ .
This gives us the partial program shown in
the internal pseudocode.
6. M-step: System recursively decomposes optimization problem. AUTO BAYES
is recursively
called with the new
goal
9 ;9 = 9 > 9 ? 9
= 9 #> 9 ? 9
wrt
.
max 6,78 Pr : <
Now, the Bayesian network
decomposition
1
#
,
schema applies with
max %32 (+* 4%5)& (+*
wrt
, revealing that is independent of
, thus the optimization problem can be decomposed into two optimization subproblems:
9
@9 : 9 #> 9 ;? 9
A9 9
B9 =
=9
max Pr <
wrt > ? and max Pr :
wrt
.
while converging
for
for &
Pr &
for &
max %')& (+* ,- .0/ Pr 1 wrt
7. System
unrolls i.i.d. Avectors.
The first subgoal from the decomposition schema,
19 : 9 > 9 ;? 9
9 ;9
max Pr <
wrt > ? , can be unrolled over the independent and identically distributed vector using an index decomposition schema which moves expressions out of
loops (sums or products) when they are not dependent on the loop index. Since # and
are co-indexed, unrolling proceeds over both (also independent and identically distributed)
#9 ;9
A9 9
vectors in parallel: max DE)C FHG Pr < E : E > ?
wrt > ? .
8. System
identifies and solves Gaussian elimination
problem. The probability Pr
#
#
is atomic because parents
. It can thus be replaced by the
appropriately instantiated Gaussian density function. Because the strictly monotone IJ#K L
function can first be applied
to the objective function of the maximization, it becomes
G
> N P
A> 9 ? 9
?
PRQ < E
67S8UT VW
67S8 N wrt
max % EC FHG %5N0M FHG+O E N
. Another application of
index decomposition allows solution for the two scalars & and & . Gaussian elimination
is then used
to solve this subproblem analytically, yielding the sequence of expressions
& % )X
% & % X
% & and & % X
% '& Y &
% )X
% & .
9. System
identifies
and solves Lagrange multiplier problem. The second subgoal
B9 = 9
=9
max Pr :
wrt
can be unrolled over the i.i.d. vector # as before. The specifica
[Z
tion condition % &BX
&
creates a constrained maximization problem in the vector
which is solved by an application of the Lagrange
multiplier schema. This in turn results
in two subproblems for a single instance & and for the multiplier which are both solved
symbolically. Thus, the usual EM algorithm for Gaussian mixtures is derived.
10. System checks and optimizes pseudocode.
During the synthesis process,
AUTO BAYES accumulates a number of constraints which have to hold to ensure proper
operation of the code (e.g., absence of divide-by-zero errors). Unless these constraints can
be resolved against the model (e.g., ]\_^ ), AUTO BAYES automatically inserts run-time
checks into the code. Before finally generating C/C++
code, the pseudocode is optimized
`Z
using information from the specification (e.g., % &BX
&
) and the domain. Thus, optimizations well beyond the capability of a regular compiler can be done.
11. System translates pseudocode to real code in desired language. Finally,
AUTO BAYES converts the intermediate code into code of the desired target system. The
source code contains thorough comments detailing the mathematics implemented. A regular compiler containing generic performance optimizations not repeated by AUTO BAYES
turns the code into an executable program. A program for sampling from a mixture of
Gaussians is also produced for testing purposes.
5 Range of Capabilities
Here, we discuss 18 examples which have been successfully handled by AUTO BAYES,
ranging from simple textbook examples to sophisticated EM models and recent multinomial versions of PCA. For each entry, the table below gives a brief description, the number
of lines of the specification and synthesized C++ code (loc), and the runtime to generate the
code (in secs., measured on a 2.2GHz Linux system). Correctness was checked for these
examples using automatically-generated test data and hand-written implementations.
Bayesian textbook examples. Simple textbook examples, like Gaussian with simple
prior
, Gaussian with inverse gamma prior
have
, or Gaussian with conjugate prior
closed-form solutions. The symbolic system of AUTO BAYES can actually find these solutions and thus generate short and efficient
code. However, a slight relaxation of the prior
on (Gaussian with semi-conjugate prior, ) requires an iterative numerical solver.
Gaussians in action.
is a Gaussian change-detection model. A slight extension of
our running example, integrating several features, yields a Gaussian Bayes classifier model
. has been successfully tested on various standard benchmarks [1], e.g., the Abalone
dataset. Currently, the number of expected classes has to be given in advance.
Mixture models and EM. A wide range of -Gaussian mixture models can be handled by
AUTO BAYES, ranging from the simple 1D (
) and 2D with diagonal covariance ( )
and with (conjugate) priors on mean
to 1D models for multi-dimensional classes
or variance . Using only a slight variation in the specification, the Gaussian distribution can be replaced by other distributions (e.g., exponentials,
, for failure analysis) or
combinations (e.g., . Gaussian and Beta,
, or -Cauchy and Poisson
). In the algorithm generated by , the analytic subsolution for the Gaussian case is combined with the
numerical solver. Finally, is a
-Gaussians and -Gaussians two-level hierarchical
mixture model which is solved by a nested instantiation of EM [15]: i.e., the M-step of the
outer EM algorithm is a second EM algorithm nested inside.
Mixtures for Regression. We represented regression with Gaussian error and Legendre
polynomials with full conjugate priors allowing smoothing [10]. Two versions of this were
then done: robust linear regression
replaces the Gaussian error with a mixture of two
Gaussians (one broad, one peaked) both centered at zero. Trajectory clustering replaces
the single regression curve by a mixture of several curves [5]. In both cases an EM algorithm is correctly integrated with the exact regression solutions.
Principal Component Analysis. We also tested a multinomial version of PCA called latent
Dirichlet allocation [2]. AUTO BAYES currently lacks variational support, yet it manages to
combine a -means style outer loop on the component proportions with an EM-style inner
loop on the hidden counts, producing the original algorithm of Hofmann, Lee and Seung,
and others [4].
#
,+
3 G
4 G
46+
4
4:9
46<
=
=
G
P
12/137
?
Q.-
>
N > G / $
16/188
P
#"*P $ &2( )"%P $ ?
?
Gauss step-detect
19/662
5 -Gauss mix 1D
17/418
Description
>
N >
P
G
???, multi-dim
???, ? prior
Gauss/Beta mix
G , P -Gauss hierarch
rob. lin. regression
mixture regression
5 5
loc
24/900
21/442
22/834
29/1053
54/1877
53/1282
0.2
0.4
2.0
0.7
1.1
0.9
1.7
2.3
14.5
9.8
#
Description
P
,0
3 P
4
470P
478
46;
G
>
! G #"%P'$ & ( )"*P $ ?
>1
N > G
P
#"%P $ & ( )"*P $ ?
?
Gauss Bayes Classify
5 -Gauss mix 2D, diag
??? 1D, > prior
5
mix
5 -Exp
-Cauchy/Poisson
mix
PCA mult/w 5 -means
? P
loc
13/148
17/233
0.2
0.4
58/1598
22/599
25/456
15/347
21/747
4.7
1.2
1.0
0.5
1.0
26/390
1.2
6 Conclusion
Beyond existing systems. Code libraries are common in statistics and learning, but they
lack the high level of automation achievable only by deep symbolic reasoning. The Bayes
Net Toolbox [12] is a Matlab library which allows users to program in models but does not
derive algorithms or generate code. The B UGS system [14] also allows users to program
in models but is specialized for Gibbs sampling. Stochastic parametrized grammars [11]
allow a concise model specification similar to AUTO BAYES ?s specification language, but
are currently only a notational device similar to XML.
Benefits of automated algorithm and code generation. Industrial-strength code. Code
generated by AUTO BAYES is efficient, validated, and commented. Extreme applications.
Extremely complex or critical applications such as spacecraft challenge the reliability limits of human-developed software. Automatically generated software allows for pervasive
condition checking and correctness-by-construction. Fast prototyping and experimentation. For both the data analyst and machine learning researcher, AUTO BAYES can function
as a powerful experimental workbench. New complex algorithms. Even with only the few
elements implemented so far, we showed that algorithms approaching research-level results
[4, 5, 10, 15] can be automatically derived. As more distributions, optimization methods
and generalized learning algorithms are added to the system, an exponentially growing
number of complex new algorithms become possible, including non-trivial variants which
may challenge any single researcher?s particular algorithm design expertise.
Future agenda. The ultimate goal is to give researchers the ability to experiment with the
entire space of complex models and state-of-the-art statistical algorithms, and to allow new
algorithmic ideas, as they appear, to be implicitly generalized to every model and special
case known to be applicable. We have already begun work on generalizing the EM schema
to continuous hidden variables, as well as adding schemas for variational methods, fast
kd-tree and -body algorithms, MCMC, and temporal models.
Availability. A web interface for AUTO BAYES is currently under development. More
information is available at http://ase.arc.nasa.gov/autobayes.
References
[1] C.L. Blake and C.J. Merz. UCI repository of machine learning databases, 1998.
[2] D. Blei, A.Y. Ng, and M. Jordan. Latent Dirichlet allocation. NIPS*14, 2002.
[3] W.L. Buntine. Operations for learning with graphical models. JAIR, 2:159?225, 1994.
[4] W.L. Buntine. Variational extensions to EM and multinomial PCA. ECML 2002, pp. 23?34, 2002.
[5] G.S. Gaffney and P. Smyth. Trajectory clustering using mixtures of regression models. In 5th KDD, pp. 63?72, 1999.
[6] Z. Ghahramani and M.J. Beal. Propagation algorithms for variational Bayesian learning. In NIPS*12, pp. 507?513, 2000.
[7] P. Haddawy. Generating Bayesian Networks from Probability Logic Knowledge Bases. In UAI 10, pp. 262?269, 1994.
[8] F. R. Kschischang, B. Frey, and H.-A. Loeliger. Factor graphs and the sum-product algorithm. IEEE Trans. Inform. Theory,
47(2):498?519, 2001.
[9] S.L. Lauritzen, A.P. Dawid, B.N. Larsen, and H.-G. Leimer. Independence properties of directed Markov fields. Networks,
20:491?505, 1990.
[10] D.J.C. Mackay. Bayesian interpolation. Neural Computation, 4(3):415?447, 1991.
[11] E. Mjolsness and M. Turmon. Stochastic parameterized grammars for Bayesian model composition. In NIPS*2000 Workshop on Software Support for Bayesian Analysis Systems, Breckenridge, December 2000.
[12] K. Murphy. Bayes Net Toolbox for Matlab. Interface of Computing Science and Statistics 33, 2001.
[13] P. Smyth, D. Heckerman, and M. Jordan. Probabilistic independence networks for hidden Markov models. Neural Computation, 9(2):227?269, 1997.
[14] A. Thomas, D.J. Spiegelhalter, and W.R. Gilks. BUGS: A program to perform Bayesian inference using Gibbs sampling.
In Bayesian Statistics 4, pp. 837?842, 1992.
[15] D.A. van Dyk. The nested EM algorithm. Statistica Sinica, 10:203-225, 2000.
| 2176 |@word repository:2 version:3 achievable:1 polynomial:1 proportion:1 seek:1 decomposition:11 simplifying:1 covariance:1 concise:1 recursively:3 functor:2 necessity:1 substitution:1 contains:2 fragment:4 loc:3 initial:6 loeliger:1 document:1 reinvented:1 existing:2 current:1 yet:1 written:2 must:2 readily:1 riacs:1 numerical:3 kdd:1 hofmann:1 enables:1 analytic:1 avectors:1 alone:1 device:1 core:3 short:1 blei:1 provides:1 node:1 ames:1 guard:2 along:2 constructed:1 beta:2 become:1 combine:2 inside:1 spacecraft:1 expected:1 roughly:1 growing:1 multi:2 ming:1 decomposed:4 automatically:7 gov:2 schumann:2 solver:2 unrolling:1 becomes:1 begin:1 moreover:3 underlying:2 lowest:1 what:2 kind:2 unspecified:1 textbook:5 developed:4 transformation:4 differentiation:2 temporal:1 thorough:1 every:1 runtime:1 classifier:1 internally:3 appear:1 producing:1 before:4 understood:1 engineering:1 frey:1 limit:1 encoding:1 accumulates:1 interpolation:1 might:1 initialization:1 quantified:1 co:1 range:2 directed:2 unique:2 gilks:1 testing:2 atomic:3 recursive:2 block:1 implement:1 procedure:1 axiom:2 thought:1 revealing:1 mult:1 integrating:1 regular:2 symbolic:11 cannot:1 selection:1 context:2 applying:1 impossible:1 optimize:6 imposed:1 rule:1 deriving:1 proving:5 variation:2 feel:1 target:2 trigger:1 construction:1 user:6 exact:2 programming:1 smyth:2 us:4 element:3 dawid:1 satisfying:3 continues:1 database:1 inserted:1 subproblem:2 solved:3 precondition:2 calculate:1 keyword:1 mjolsness:1 observes:1 mu:4 seung:1 dom:1 ultimately:1 rewrite:1 algebra:4 creation:1 creates:1 completely:1 easily:2 resolved:1 represented:5 various:2 derivation:8 instantiated:1 fast:2 describe:3 query:1 detected:1 whose:1 widely:2 solve:2 turmon:1 drawing:1 grammar:2 ability:2 statistic:3 fischer:1 syntactic:1 a9:2 beal:1 sequence:1 net:3 product:4 maximal:1 uci:1 combining:1 loop:8 iff:2 fisch:1 flexibility:1 achieve:1 bug:3 description:3 recipe:1 parent:6 double:5 extending:1 produce:1 generating:2 derive:2 measured:1 lauritzen:2 keywords:1 solves:2 implemented:5 c:1 involves:1 implies:1 functionality:1 stochastic:2 centered:1 human:1 elimination:2 implementing:2 require:4 inventing:1 decompose:2 elementary:1 summation:4 extension:5 strictly:1 subprogram:1 hold:7 insert:1 b9:2 ground:1 blake:1 exp:1 algorithmic:2 automate:1 vary:1 purpose:1 compiles:1 applicable:1 combinatorial:1 label:1 currently:6 grind:1 correctness:2 successfully:2 tool:2 gaussian:18 always:1 pervasive:1 encode:1 derived:6 focus:1 validated:1 notational:1 likelihood:2 check:2 industrial:1 detect:1 posteriori:2 inference:4 dim:1 dependent:1 entire:2 integrated:1 hidden:6 deduction:1 ancestor:5 transformed:2 fhg:3 overall:1 html:1 denoted:1 development:1 smoothing:1 art:2 integration:1 initialize:1 constrained:1 equal:2 construct:2 never:1 special:1 ng:1 sampling:7 manually:1 eliminated:1 field:1 broad:1 look:1 peaked:1 future:1 simplex:1 others:2 few:1 unsound:1 gamma:1 tightly:2 murphy:1 replaced:3 intended:1 phase:2 attempt:2 detection:1 interest:1 gaffney:1 highly:2 custom:1 evaluation:1 mixture:14 navigation:1 extreme:1 yielding:1 johann:1 amenable:1 integral:1 closer:1 partial:1 necessary:1 respective:1 unless:1 indexed:6 tree:1 loosely:1 unite:1 divide:1 desired:2 detailing:1 minimal:2 instance:4 classify:1 earlier:1 assertion:2 maximization:2 applicability:4 cost:1 subset:1 imperative:1 entry:1 buntine:4 ugs:1 dependency:2 answer:1 synthetic:1 combined:1 density:1 fundamental:2 ancestral:1 minimality:1 probabilistic:3 systematic:1 lee:1 synthesis:7 linux:1 opposed:1 containing:1 creating:1 style:2 return:1 prolog:1 converted:2 de:1 sec:1 automation:1 int:3 availability:1 descendent:4 explicitly:1 piece:1 tion:1 performed:2 root:1 closed:3 break:1 schema:32 linked:1 reached:1 compiler:5 bayes:29 start:1 capability:5 parallel:1 annotation:1 variance:2 efficiently:2 correspond:1 yield:2 conceptually:1 generalize:1 bayesian:22 wray:1 disc:2 backtrack:1 produced:2 trajectory:2 manages:1 researcher:3 expertise:1 inform:1 unrolls:1 email:1 checked:1 against:1 failure:1 mathematica:2 involved:1 pp:5 larsen:1 con:1 gain:1 dataset:2 popular:1 begun:1 knowledge:2 formalize:1 routine:1 sophisticated:1 actually:2 nasa:3 jair:1 follow:1 specify:2 formulation:1 hiit:1 done:6 evaluated:1 subgoal:2 until:1 working:1 hand:1 web:1 lack:3 propagation:1 defines:1 gray:1 perhaps:1 building:1 effect:1 contain:2 multiplier:3 analytically:1 during:3 anything:1 abalone:1 criterion:2 generalized:2 octave:1 plate:1 outline:1 performs:1 interface:2 reasoning:2 ranging:3 meld:1 variational:5 novel:1 fi:1 recently:1 common:2 specialized:3 uto:1 multinomial:5 executable:2 pseudocode:4 overview:1 conditioning:1 operationalized:1 exponentially:1 s8:1 slight:3 synthesized:1 mellon:1 composition:1 gibbs:4 vec:2 breckenridge:1 automatic:4 mathematics:2 language:6 reliability:1 moving:2 specification:10 add:1 base:1 recent:2 showed:1 optimizes:1 driven:1 manipulation:1 underlined:1 additional:1 determine:1 maximize:1 semi:1 full:3 mix:5 match:3 characterized:1 lin:1 ase:1 halt:1 converging:3 involving:1 regression:9 variant:1 cmu:1 mog:1 poisson:2 sometimes:1 represent:1 mex:1 hurdle:1 annealing:1 source:1 appropriately:1 operate:1 strict:1 comment:5 file:1 facilitates:1 december:1 jordan:2 integer:1 call:2 intermediate:5 identically:3 automated:2 independence:3 marginalization:2 identified:1 approaching:1 inner:1 idea:1 translates:1 enumerating:1 expression:9 pca:6 handled:2 ultimate:1 returned:1 hessian:1 action:1 matlab:5 deep:1 processed:1 documented:1 generate:5 specifies:3 http:1 exist:1 blown:1 correctly:1 subpart:1 carnegie:1 intertwined:1 discrete:7 commented:2 hierarch:1 graph:4 symbolically:1 monotone:1 relaxation:1 sum:6 convert:2 run:1 package:1 inverse:1 powerful:2 parameterized:1 family:2 throughout:1 separation:2 datalog:1 comparable:1 simplification:4 replaces:3 strength:1 constraint:5 helsinki:1 software:5 generates:2 aspect:1 argument:2 extremely:1 performing:1 ayes:1 creative:1 combination:2 conjugate:5 legendre:1 smaller:1 across:2 em:24 describes:1 kd:1 heckerman:1 rob:1 pr:16 computationally:1 visualization:1 turn:2 discus:1 mechanism:1 count:1 wrt:13 needed:1 end:1 serf:1 available:2 gaussians:6 operation:2 experimentation:1 apply:1 hierarchical:1 away:1 generic:4 existence:3 original:2 thomas:1 standardized:1 clustering:3 running:2 include:2 ensure:1 graphical:2 dirichlet:2 readable:1 const:2 invokes:1 ghahramani:1 workbench:1 move:2 objective:2 question:2 prover:2 already:3 added:2 quantity:2 strategy:1 usual:1 nr:2 prq:1 diagonal:1 gradient:1 separate:1 simulated:1 parametrized:1 outer:2 me:1 cauchy:2 trivial:1 analyst:1 code:34 index:5 ratio:1 unrolled:2 sinica:1 executed:1 inwards:1 statement:2 potentially:1 subproblems:6 sigma:4 negative:1 disparate:1 design:8 implementation:3 proper:1 agenda:1 unknown:1 perform:2 allowing:3 markov:2 arc:2 benchmark:1 ecml:1 immediate:1 situation:1 extended:2 introduced:2 bernd:1 mechanical:1 toolbox:4 required:2 optimized:2 specified:2 engine:1 nip:3 trans:1 beyond:5 bar:1 proceeds:1 below:2 usually:2 prototyping:1 reading:1 challenge:3 program:15 max:13 including:2 explanation:1 critical:1 sian:1 customized:1 scheme:2 xml:1 spiegelhalter:1 brief:1 library:2 identifies:3 ready:1 auto:24 extract:1 coupled:1 dyk:1 text:1 prior:9 understanding:1 checking:1 subexpressions:1 par:1 mixed:1 generation:3 rameters:1 allocation:2 proven:1 principle:1 last:1 free:1 side:1 allow:3 institute:1 fall:1 wide:1 distributed:2 ghz:1 curve:2 benefit:1 van:1 numeric:1 stand:1 forward:1 collection:1 commonly:1 universally:1 simplified:1 far:2 compact:1 implicitly:2 logic:5 dealing:1 instantiation:1 uai:1 summing:2 assumed:1 freshly:1 continuous:1 latent:4 iterative:1 vergence:1 decomposes:1 table:1 nature:1 robust:1 inherently:1 kschischang:1 improving:1 agray:1 complex:7 domain:2 diag:1 main:1 statistica:1 repeated:2 body:1 crafted:1 fashion:1 formalization:1 theme:1 wish:1 exponential:1 jacobian:1 externally:1 down:2 theorem:8 formula:6 specific:2 symbol:2 incorporating:1 exists:1 workshop:1 adding:2 suited:1 generalizing:1 logarithmic:1 lagrange:2 phi:5 scalar:1 applies:2 radically:1 nested:3 mackay:1 extracted:1 conditional:2 goal:8 towards:1 replace:2 absence:1 feasible:2 change:2 infinite:1 lemma:12 principal:2 called:2 gauss:8 experimental:1 merz:1 internal:1 support:3 alexander:1 constructive:1 evaluate:2 mcmc:1 tested:2 scratch:1 handling:1 |
1,292 | 2,177 | Learning Sparse Topographic Representations
with Products of Student-t Distributions
Max Welling and Geoffrey Hinton
Department of Computer Science
University of Toronto
10 King?s College Road
Toronto, M5S 3G5 Canada
welling,hinton @cs.toronto.edu
Simon Osindero
Gatsby Unit
University College London
17 Queen Square
London WC1N 3AR, UK
[email protected]
Abstract
We propose a model for natural images in which the probability of an image is proportional to the product of the probabilities of some filter outputs. We encourage the system to find sparse features by using a Studentt distribution to model each filter output. If the t-distribution is used to
model the combined outputs of sets of neurally adjacent filters, the system learns a topographic map in which the orientation, spatial frequency
and location of the filters change smoothly across the map. Even though
maximum likelihood learning is intractable in our model, the product
form allows a relatively efficient learning procedure that works well even
for highly overcomplete sets of filters. Once the model has been learned
it can be used as a prior to derive the ?iterated Wiener filter? for the purpose of denoising images.
1 Introduction
Historically, two different classes of statistical model have been used for natural images.
?Energy-based? models assign to each image a global energy, , that is the sum of a number of local contributions and they define the probability of an image to be proportional to
. This class of models includes Markov Random Fields where combinations of
nearby pixel values contribute local energies, Boltzmann Machines in which binary pixels
are augmented with binary hidden variables that learn to model higher-order statistical interactions and Maximum Entropy methods which learn the appropriate magnitudes for the
energy contributions of heuristically derived features [5] [9]. It is difficult to perform maximum likelihood fitting on most energy-based models because of the normalization term
(the partition function) that is required to convert
to a probability. The normalization term is a sum over all possible images and its derivative w.r.t. the parameters is
required for maximum likelihood fitting. The usual approach is to approximate this derivative by using Markov Chain Monte Carlo (MCMC) to sample from the model, but the large
number of iterations required to reach equilibrium makes learning very slow.
The other class of model uses a ?causal? directed acyclic graph in which the lowest level
nodes correspond to pixels and the probability distribution at a node (in the absence of
any observations) depends only on its parents. When the graph is singly or very sparsely
connected there are efficient algorithms for maximum likelihood fitting but if nodes have
many parents, it is hard to perform maximum likelihood fitting because this requires the
intractable posterior distribution over non-leaf nodes given the pixel values.
There is much debate about which class of model is the most appropriate for natural images.
Is a particular image best characterized by the states of some hidden variables in a causal
generative model? Or is it best characterized by its peculiarities i.e. by saying which of a
very large set of normally satisfied constraints are violated? In this paper we treat violations
of constraints as contributions to a global energy and we show how to learn a large set of
constraints each of which is normally satisfied fairly accurately but occasionally violated
by a lot. The ability to learn efficiently without ever having to generate equilibrium samples
from the model and without having to confront the intractable partition function removes a
major obstacle to the use of energy-based models.
2 The Product of Student-t Model
Products of Experts (PoE) are a restricted class of energy-based model [1]. The distribution
represented by a PoE is simply the normalized product of all the distributions represented
by the individual ?experts?:
(1)
where are un-normalized experts and denotes the overall normalization constant.
In the product
of Student-t (PoT) model, un-normalized experts have the following form,
"!
(2)
where is called a filter and is the # -th column in the filter-matrix . When properly
normalized, this represents a Student-t distribution over the filtered random variable $
. An important feature of the Student-t distribution is its heavy tails, which makes it a
suitable candidate for modelling constraints of the kind that are found in images.
Defining
%
%
&
%
, the energy of the PoT model becomes
'
(
)+*-,
.
(3)
Viewed this way, the model takes the form of a maximum entropy distribution with weights
on real-valued ?features? of the image. Unlike previous maximum entropy models,
however, we can fit both the weights and the features at the same time.
When the number of input dimensions is equal to the number of experts, the normally intractable partition function becomes a determinant and the PoT model becomes equivalent
to a noiseless ICA model with Student-t prior distributions [2]. In that case the rows of
the inverse filters /0 21 will represent independent directions in input space. So noiseless ICA can be viewed as an energy-based model even though it is usually interpeted as
a causal generative model in which the posterior over the hidden variables collapses to a
point. However, when we consider more experts than input dimensions (i.e. an overcomplete representation), the energy-based view and the causal generative view lead to different
generalizations of ICA. The natural causal generalization retains the independence of the
hidden variables in the prior by assuming independent sources. In contrast, the PoT model
simply multiplies together more experts than input dimensions and re-normalizes to get the
total probability.
3 Training the PoT Model with Contrastive Divergence
When training energy-based models we need to shape the energy function so that observed
images have low energy and empty regions in the space of all possible images have high
energy. The maximum likelihood learning rule is given by,
(4)
It is the second term which causes learning to be slow and noisy because it is usually
necessary to use MCMC to compute the average over the equilibrium distribution. A much
more efficient way to fit the model is to use the data distribution itself to initialize a Markov
Chain which then starts moving towards the model?s equilibrium distribution. After just a
few steps, we observe how the chain is diverging from the data and adjust the parameters
to counteract this divergence. This is done by lowering the energy of the data and raising
the energy of the ?confabulations? produced by a few steps of MCMC.
#"
$%"&'
!
(5)
1
It can be shown that the above update rule approximately minimizes a new objective function called the contrastive divergence [1].
As it stands the learning rule will be inefficient if the Markov Chain mixes slowly because
the two terms in equation 5 will almost cancel each other out. To speed up learning we need
a Markov chain that mixes rapidly so that the confabulations will be some distance away
from the data. Rapid mixing can be achieved by alternately Gibbs sampling a set of hidden
variables given the random variables under consideration and vice versa. Fortunately, the
PoT model can be equipped with a number of hidden random variables equal to the number
of experts as follows,
)(#* ,+ %-/1. 0&2
3 4 5 76 82 :5 9=; <> 8 > 6 5
Integrating over the *
1
1
>@? ACB 4 ED
(6)
variables results in the density of the PoT model, i.e. eqns. (1) and
(2). Moreover, the conditional distributions are easy to identify and sample from, namely
M
LK
(7)
J
I
HG 4
S Q
*
(8)
N <PO K RQ
UTPVFWYX 3 Z D
where denotes a Gamma distribution and N a normal distribution. From (8) we see that
* can be interpreted as precision variables in the transformed space [
G
.
the variables
F*
'
!
.
1
In this respect our model resembles a ?Gaussian scale mixture? (GSM) [8] which also
multiplies a positive scaling variable with a normal variate. But GSM is a causal model
while PoT is energy-based.
The (in)dependency relations between the variables in a PoT model are depicted graphically
in figure (1a,b). The hidden variables are independent given , which allows them to be
Gibbs-sampled in parallel. This resembles the way in which brief Gibbs sampling is used
to fit binary ?Restricted Boltzmann Machines? [1].
To learn the parameters of the PoT model we thus propose to iterate the following steps:
*^\ ]
1) Sample
given the data
distribution (7).
_]
for every data-vector according to the Gamma-
u
u
u
T 2
(Jx)
J
x
x
(a)
T 2
(Jx)
1
2
3
4
5
W
J
x
(b)
(c)
(d)
Figure 1: (a)- Undirected graph for the PoT model. (b)-Expanded graph where the
deterministic
relation (dashed lines) between the random variable and the activities of the filters
is made
explicit. (c)-Graph for the PoT model including weights . (d)-Filters with large (decreasing from
left to right) weights into a particular top level unit . Top level units have learned to connect to
filters similar in frequency, location and orientation.
\]
2) Sample reconstructions of the data given the sampled values of
data-vector according to the Normal distribution (8).
*\ ]
for every
R\ ]
3) Update the parameters according to (5) where the ?k-step samples? are now given
by the reconstructions , the energy is given by (3), and the parameters are given
by
.
)(
4 Overcomplete Representations
The above learning rules are still valid for overcomplete representations. However, step-2
of the learning algorithm is much more efficient when the inverse of the filter matrix
exists. In that case we simply draw standard normal random numbers (with the num1
ber of data-vectors) and multiply each of them with 1
. This is efficient because
1
is diagonal while the costly inverse 1 is data indepenthe data dependent matrix
dent. In contrast, for the overcomplete case we ought to perform a Cholesky factorization
on
for each data-vector separately. We have, however, obtained good results by
proceeding as in the complete case and replacing the inverse of the filter matrix with its
pseudo-inverse.
Q ] 82
Q ] 82
RQP]
From experiments we have also
found that in the overcomplete case we should fix the
norm of the filters,
# , in order to prevent some of them from decaying to zero.
This operation is done after every step of learning. Since controlling the norm removes the
ability of the experts to adapt to scale it is necessary to whiten the data first.
4.1 Experiment: Overcomplete Representations for Natural Images
(
We randomly generated ! !!! patches of ! ! pixels from images of natural scenes 1 .
The patches were centered and sphered using PCA and the DC component (eigen-vector
with largest variance) was removed. The algorithm
for overcomplete representations using
"!
the pseudo-inverse was used to train !
experts, i.e. a representation
that is more than
!-!
times overcomplete.
We
fixed
the
weights
to
have
and
the
the
filters to have
a -norm of . A small weight decay term and a momentum
term were included in the
gradient updates of the filters. The learning rate was set so that initially the change in the
filters was approximately !# !! . In figure (2a) we show a small subset
of the inverse-filters
given by the pseudo-inverse of %$'&)(%* , where $+&)(%* is the !!,.-/- matrix used for
sphering the data.
(
1
Collected from http://www.cis.hut.fi/projects/ica/data/images
5 Topographically Ordered Features
In [6] it was shown that linear filtering of natural images is not enough to remove all higher
order dependencies.
In particular,
it was argued that there are residual dependencies among
the activities $ of the filtered inputs. It is therefore desirable to model those
dependencies within the PoT model. By inspection of figure (1b) we note that these dependencies can be modelled through a non-negative weight matrix ! , which connects
the hidden variables with the activities . The resultant model is depicted in figure
(1c).
Depending on how many nonzero weights emanate from a hidden unit (say
), each expert now occupies input dimensions instead of justone.
The expressions
for
these richer experts can be obtained from (2) by replacing,
. We
# ).
have found that learning is assisted by fixing the -norm of the weights (
Moreover, we have found that the sparsity of the weights can be controlled by the following generalization of the experts,
Z
Z
-
(
"!
(
!
(9)
The larger the value for the sparser the distribution of
values.
Joint and conditional distributions over hidden variables are obtained through similar replacements in eqn. (6) and (7) respectively. Sampling the reconstructions given the states of
the hidden variables proceeds by first sampling from independent generalized
Laplace
which are
distributions
with precision parameters 0
subsequently transformed into 1 . Learning in this model therefore proceeds with
only minor modifications to the algorithm described in the previous section.
*
\
When we learn the weight matrix from image data we find that a particular hidden
variable develops weights to the activities of filters similar in frequency, location and
orientation. The variables therefore integrate information from these filters and as a
result develop certain invariances that resemble the behavior of complex cells. A similar
approach was studied in [4] using a related causal model 2 in which a number of scale
variables generate correlated variances for conditionally Gaussian experts. This results in
topography when the scale-generating variables are non-adaptive and connect to a local
neighborhood of filters only.
Z
*
We will now argue that fixed local weights
also give rise to topography in the PoT
model. The reason is that averaging the squares of randomly chosen filter outputs (eqn.9)
produces an approximately Gaussian distribution which is a poor fit to the heavy-tailed
experts. However, this ?smoothing effect? may be largely avoided by averaging squared
filter outputs that are highly correlated (i.e. ones that are similar in location, frequency and
orientation). Since the averaging is local, this results in a topographic layout of the filters.
5.1 Experiment: Topographic Representations for Natural Images
(
!
!
pixels in the
same
For this experiment we collected ! !!-! image patches of size
!
way as described in section
(4.1).
The
image
data
were
sphered
and
reduced
to
di!
mensions by removing low variance and high variance (DC) direction. We learned
an
.
.
overcomplete representation with !! experts which were organized on a square !, !
&
grid. Each expert connects with a fixed weight of
to itself and all its neighbors, where periodic boundary conditions were imposed for the experts on the boundary.
2
Interestingly, the update equations for the filters presented in [4], which minimize a bound on
the log-likelihood of a directed model, reduce to the same equations as our learning rules when the
representation is complete and the filters orthogonal.
(a)
(b)
Figure 2: (a)-Small subset of the learned filters from a
times overcomplete representation for natural image patches. (b)-Topographically ordered filters. The weights were fixed and
connect to neighbors only, using periodic boundary conditions. Neighboring filters have learned to
be similar in frequency, location and orientation. One can observe a pinwheel structure to the left of
the low frequency cluster.
.
We adapted the filters ( -norm ) and used fixed values for
and
. The resulting inverse-filters are shown in figure (2b). We note that the weights have enforced a
topographic ordering on the experts, where location, scale and frequency of the Gabor-like
filters all change smoothly across the map.
!
In another experiment we used the same data to train
a complete representation of
experts where we learned the weights ( -norm ), and the filters (unconstrained),
but with a fixed value of
. The weights and were kept positive by adapting their
can now connect to any other expert we do not expect
logarithm. Since the weights
topography. To study whether the weights
were modelling the dependencies between
the energies of the filter outputs we ordered the filters for each complex cell
according to the strength of the weights connecting to it. For a representative subset of the
complex cells , we show the ! filters with the strongest connections to that cell in figure
(1d). Since the cells connect to similar filters we may conclude that the weights
are
indeed learning the dependencies between the activities of the filter outputs.
Z
*
6 Denoising Images: The Iterated Wiener Filter
If the PoT model provides an accurate description of the statistics of natural image data it
ought to be a good prior for cleaning up noisy images. In the following we will apply this
idea to denoise images contaminated with Gaussian pixel noise. We follow the standard
Bayesian approach which states that the optimal estimate of the original image is given by
the maximum a posteriori (MAP) estimate of , where denotes the noisy image.
For the PoT model this reduces to,
*
&
W X V <
.
1
(
)*,
(
.
(10)
(a)
(b)
(c)
(d)
Figure 3: (a)- Original ?rock?-image. (b)-Rock-image with noise added. (c)-Denoised image using
Wiener filtering. (d) Denoised image using IWF.
To minimize this we follow a variational procedure where we upper bound the logarithm
%
%
)+*-,
&
using )+*-, %
. The bound is saturated when
. Applying this to
every logarithm in the summation in eqn. (10) and iteratively minimizing this bound over
and we find the following update equations,
*
&
&
(
.
1
T
1
1
(11)
T UTPVFWYX 3
0
D
(12)
where denotes componentwise multiplication. Since the second equation is just a Wiener
filter with noise covariance and a Gaussian prior with covariance 1 we have
named the above denoising equations the iterated Wiener filter (IWF).
T
When the filters are orthonormal, the noise covariance isotropic and the weight matrix the
identity, the minimization in (10) decouples into minimizations over the transformed
variables $ . Defining we can easily derive that $ is the solution of the
following cubic equation (for which analytic solutions exist),
$
$
.
$
.
!
(13)
We note however that constraining the filters to be orthogonal is a rather severe restriction if
the data are not pre-whitened. On the other hand, if we decide to work with whitened data,
the isotropic noise assumption seems unrealistic. Having said that, Hyvarinen?s shrinkage
method for ICA models [3] is based on precisely these assumptions and seems to give good
results. The proposed method is also related to approaches based on the GSM [7].
6.1 Experiment: Denoising
To test the iterated Wiener filter, we trained a complete set of - - experts on the data described in section (4.1). The norm of the filters was unconstrained, the were free to adapt,
but we did not include any weights . The image
shown in figure (3a) was corrupted
with
.
..
!
Gaussian noise with standard deviation
, which resulted in a PSNR of # ! dB (figure (3b)). We applied the adaptive Wiener filter from matlab (Wiener2.m) with an optimal
neighborhood size and known noise-variance. The denoised image using adaptive
."!
Wiener filtering
has a PSNR of # - dB and is shown in figure (3c). IWF was run on every
possible ! ! patch in the image, after which the results were averaged. Because the
filters were trained on sphered data without a DC component, the same transformations
have to be applied to the test patches before
IWF
is applied. The denoised image using
.
#
dB,
which is a significant improvement of
IWF
is
shown
in
(3d)
and
has
a
PSNR
of
.
#
dB over Wiener filtering. It is our hope that the use of overcomplete representations and
weights will further improve those results.
7 Discussion
It is well known that a wavelet transform de-correlates natural image data in good approximation. In [6] it was found that in the marginal distribution the wavelet coefficients are
sparsely
distributed but that there are significant residual dependencies among their ener
gies $ . In this paper we have shown that the PoT model can learn highly overcomplete filters with sparsely distributed outputs. With a second hidden layer that is locally connected,
it captures the dependencies between filter outputs by learning topographic representations.
Our approach improves upon earlier attempts (e.g. [4],[8]) in a number of ways. In the
PoT model the hidden variables are conditionally independent so perceptual inference is
very easy and does not require iterative settling even when the model is overcomplete.
There is a fairly simple and efficient procedure for learning all the parameters, including
the weights connecting top-level units to filter outputs. Finally, the model leads to an
elegant denoising algorithm which involves iterating a Wiener-filter.
Acknowledgements
This research was funded by NSERC, the Gatsby Charitable Foundation, and the Wellcome Trust.
We thank Yee-Whye Teh for first suggesting a related model and Peter Dayan for encouraging us to
apply products of experts to topography.
References
[1] G.E. Hinton. Training products of experts by minimizing contrastive divergence. Neural Computation, 14:1771?1800, 2002.
[2] G.E. Hinton, M. Welling, Y.W. Teh, and K. Osindero. A new view of ICA. In Int. Conf. on
Independent Component Analysis and Blind Source Separation, 2001.
[3] A. Hyvarinen. Sparse code shrinkage: Denoising of nongaussian data by maximum likelihood
estimation. Neural Computation, 11(7):1739?1768, 1999.
[4] A. Hyvarinen, P.O. Hoyer, and M. Inki. Topographic independent component analysis. Neural
Computation, 13(7):1525?1558, 2001.
[5] S. Della Pietra, V.J. Della Pietra, and J.D. Lafferty. Inducing features of random fields. IEEE
Transactions on Pattern Analysis and Machine Intelligence, 19(4):380?393, 1997.
[6] E.P. Simoncelli. Modeling the joint statistics of images in the wavelet domain. In Proc SPIE,
44th Annual Meeting, volume 3813, pages 188?195, Denver, 1999.
[7] V. Strela, J. Portilla, and E. Simoncelli. Image denoising using a local Gaussian scale mixture
model in the wavelet domain. In Proc. SPIE, 45th Annual Meeting, San Diego, 2000.
[8] M.J. Wainwright and E.P. Simoncelli. Scale mixtures of Gaussians and the statistics of natural
images. In Advances Neural Information Processing Systems, volume 12, pages 855?861, 2000.
[9] S.C. Zhu, Z.N. Wu, and D. Mumford. Minimax entropy principle and its application to texture
modeling. Neural Computation, 9(8):1627?1660, 1997.
| 2177 |@word determinant:1 norm:7 seems:2 heuristically:1 covariance:3 contrastive:3 interestingly:1 partition:3 shape:1 analytic:1 remove:3 update:5 generative:3 leaf:1 intelligence:1 inspection:1 isotropic:2 filtered:2 provides:1 contribute:1 toronto:3 location:6 node:4 fitting:4 wiener2:1 indeed:1 ica:6 rapid:1 behavior:1 decreasing:1 encouraging:1 equipped:1 becomes:3 project:1 moreover:2 lowest:1 strela:1 kind:1 interpreted:1 minimizes:1 transformation:1 ought:2 pseudo:3 every:5 decouples:1 uk:2 unit:5 normally:3 positive:2 before:1 local:6 treat:1 approximately:3 resembles:2 studied:1 collapse:1 factorization:1 averaged:1 directed:2 procedure:3 emanate:1 gabor:1 adapting:1 pre:1 road:1 integrating:1 get:1 applying:1 yee:1 www:1 equivalent:1 map:4 deterministic:1 imposed:1 restriction:1 graphically:1 layout:1 rule:5 orthonormal:1 laplace:1 controlling:1 diego:1 cleaning:1 us:1 sparsely:3 observed:1 capture:1 region:1 connected:2 ordering:1 removed:1 rq:1 trained:2 topographically:2 upon:1 po:1 joint:2 easily:1 represented:2 train:2 london:2 monte:1 neighborhood:2 richer:1 larger:1 valued:1 say:1 ability:2 statistic:3 topographic:7 transform:1 noisy:3 itself:2 rock:2 ucl:1 propose:2 reconstruction:3 interaction:1 product:9 neighboring:1 rapidly:1 mixing:1 description:1 inducing:1 parent:2 empty:1 cluster:1 produce:1 generating:1 derive:2 depending:1 ac:1 iwf:5 fixing:1 develop:1 minor:1 pot:18 c:1 resemble:1 involves:1 direction:2 filter:52 peculiarity:1 subsequently:1 centered:1 occupies:1 argued:1 require:1 assign:1 fix:1 generalization:3 summation:1 dent:1 assisted:1 hut:1 normal:4 equilibrium:4 major:1 jx:2 purpose:1 estimation:1 proc:2 largest:1 vice:1 minimization:2 hope:1 gaussian:7 rather:1 shrinkage:2 derived:1 properly:1 improvement:1 modelling:2 likelihood:8 contrast:2 posteriori:1 inference:1 dependent:1 dayan:1 initially:1 hidden:14 relation:2 transformed:3 pixel:7 overall:1 among:2 orientation:5 multiplies:2 spatial:1 smoothing:1 fairly:2 initialize:1 marginal:1 field:2 once:1 equal:2 having:3 sampling:4 represents:1 cancel:1 contaminated:1 develops:1 few:2 randomly:2 gamma:2 divergence:4 resulted:1 individual:1 pietra:2 connects:2 replacement:1 attempt:1 highly:3 multiply:1 adjust:1 saturated:1 severe:1 violation:1 mixture:3 hg:1 wc1n:1 chain:5 accurate:1 encourage:1 necessary:2 orthogonal:2 logarithm:3 re:1 causal:7 overcomplete:14 column:1 earlier:1 obstacle:1 modeling:2 ar:1 retains:1 queen:1 deviation:1 subset:3 osindero:2 acb:1 dependency:9 connect:5 periodic:2 corrupted:1 combined:1 density:1 together:1 connecting:2 nongaussian:1 squared:1 satisfied:2 slowly:1 conf:1 expert:24 derivative:2 inefficient:1 suggesting:1 de:1 gy:1 student:6 includes:1 coefficient:1 int:1 depends:1 blind:1 view:3 lot:1 start:1 decaying:1 denoised:4 parallel:1 simon:2 contribution:3 minimize:2 square:3 wiener:10 variance:5 largely:1 efficiently:1 correspond:1 identify:1 modelled:1 bayesian:1 iterated:4 accurately:1 produced:1 carlo:1 m5s:1 gsm:3 strongest:1 reach:1 ed:1 energy:20 frequency:7 resultant:1 di:1 spie:2 sampled:2 improves:1 psnr:3 organized:1 higher:2 follow:2 done:2 though:2 just:3 hand:1 eqn:3 replacing:2 trust:1 effect:1 normalized:4 nonzero:1 iteratively:1 conditionally:2 adjacent:1 eqns:1 whiten:1 generalized:1 whye:1 complete:4 image:40 variational:1 consideration:1 fi:1 inki:1 denver:1 volume:2 tail:1 significant:2 versa:1 gibbs:3 unconstrained:2 grid:1 funded:1 moving:1 posterior:2 occasionally:1 certain:1 binary:3 meeting:2 fortunately:1 dashed:1 neurally:1 mix:2 desirable:1 reduces:1 simoncelli:3 characterized:2 adapt:2 controlled:1 whitened:2 confront:1 noiseless:2 iteration:1 normalization:3 represent:1 achieved:1 cell:5 separately:1 source:2 unlike:1 elegant:1 undirected:1 db:4 lafferty:1 constraining:1 easy:2 enough:1 iterate:1 independence:1 fit:4 variate:1 reduce:1 idea:1 poe:2 expression:1 pca:1 whether:1 studentt:1 peter:1 cause:1 matlab:1 iterating:1 singly:1 locally:1 reduced:1 generate:2 http:1 exist:1 mensions:1 prevent:1 kept:1 lowering:1 graph:5 sum:2 convert:1 enforced:1 counteract:1 inverse:9 run:1 named:1 saying:1 almost:1 decide:1 wu:1 patch:6 separation:1 draw:1 scaling:1 bound:4 layer:1 annual:2 activity:5 adapted:1 strength:1 constraint:4 precisely:1 scene:1 nearby:1 speed:1 expanded:1 relatively:1 sphering:1 department:1 according:4 combination:1 poor:1 across:2 modification:1 restricted:2 ener:1 wellcome:1 equation:7 operation:1 gaussians:1 apply:2 observe:2 away:1 appropriate:2 confabulation:2 eigen:1 original:2 denotes:4 top:3 include:1 objective:1 added:1 mumford:1 costly:1 usual:1 diagonal:1 g5:1 said:1 hoyer:1 gradient:1 distance:1 thank:1 argue:1 collected:2 reason:1 assuming:1 code:1 minimizing:2 difficult:1 debate:1 negative:1 rise:1 boltzmann:2 perform:3 teh:2 upper:1 observation:1 markov:5 pinwheel:1 defining:2 hinton:4 ever:1 dc:3 portilla:1 canada:1 namely:1 required:3 connection:1 componentwise:1 raising:1 learned:6 alternately:1 proceeds:2 usually:2 pattern:1 sparsity:1 max:1 including:2 wainwright:1 unrealistic:1 suitable:1 natural:12 settling:1 residual:2 zhu:1 minimax:1 improve:1 historically:1 brief:1 lk:1 prior:5 acknowledgement:1 multiplication:1 expect:1 topography:4 proportional:2 filtering:4 acyclic:1 geoffrey:1 foundation:1 integrate:1 principle:1 charitable:1 heavy:2 row:1 normalizes:1 free:1 ber:1 neighbor:2 sparse:3 distributed:2 boundary:3 dimension:4 stand:1 valid:1 made:1 adaptive:3 san:1 avoided:1 welling:3 hyvarinen:3 correlate:1 transaction:1 approximate:1 global:2 conclude:1 un:2 iterative:1 tailed:1 learn:7 complex:3 domain:2 did:1 noise:7 denoise:1 augmented:1 representative:1 cubic:1 gatsby:3 slow:2 precision:2 momentum:1 explicit:1 candidate:1 rqp:1 perceptual:1 learns:1 wavelet:4 removing:1 decay:1 intractable:4 exists:1 ci:1 texture:1 magnitude:1 sparser:1 smoothly:2 entropy:4 depicted:2 simply:3 ordered:3 nserc:1 conditional:2 viewed:2 identity:1 king:1 towards:1 absence:1 change:3 hard:1 included:1 averaging:3 denoising:7 called:2 total:1 invariance:1 diverging:1 college:2 cholesky:1 sphered:3 violated:2 mcmc:3 della:2 correlated:2 |
1,293 | 2,178 | Neural Decoding of Cursor Motion Using a Kalman Filter
W. Wu
M. J. Black
Y. Gao
M. Serruya
A. Shaikhouni
E. Bienenstock
J. P. Donoghue
Division of Applied Mathematics, Dept. of Computer Science,
Dept. of Neuroscience, Division of Biology and Medicine,
Brown University, Providence, RI 02912
[email protected], [email protected], [email protected],
[email protected], Mijail [email protected],
Ammar [email protected], john [email protected]
Abstract
The direct neural control of external devices such as computer displays
or prosthetic limbs requires the accurate decoding of neural activity representing continuous movement. We develop a real-time control system
using the spiking activity of approximately 40 neurons recorded with
an electrode array implanted in the arm area of primary motor cortex.
In contrast to previous work, we develop a control-theoretic approach
that explicitly models the motion of the hand and the probabilistic relationship between this motion and the mean firing rates of the cells in
70 bins. We focus on a realistic cursor control task in which the subject must move a cursor to ?hit? randomly placed targets on a computer
monitor. Encoding and decoding of the neural data is achieved with a
Kalman filter which has a number of advantages over previous linear
filtering techniques. In particular, the Kalman filter reconstructions of
hand trajectories in off-line experiments are more accurate than previously reported results and the model provides insights into the nature of
the neural coding of movement.
1 Introduction
Recent results have demonstrated the feasibility of direct neural control of devices such as
computer cursors using implanted electrodes [5, 9, 11, 14]. These results are enabled by a
variety of mathematical ?decoding? methods that produce an estimate of the system ?state?
(e.g. hand position) from a sequence of measurements (e.g. the firing rates of a collection
of cells). Here we argue that such a decoding method should (1) have a sound probabilistic
foundation; (2) explicitly model noise in the data; (3) indicate the uncertainty in estimates
of hand position; (4) make minimal assumptions about the data; (5) require a minimal
amount of ?training? data; (6) provide on-line estimates of hand position with short delay
(less than 200ms); and (7) provide insight into the neural coding of movement. To that
Monitor
12
Target
10
8
6
Tablet
Trajectory
4
2
Manipulandum
a
2
4
6
8
10
12
14
16
b
Figure 1: Reconstructing 2D hand motion. (a) Training: neural spiking activity is recorded
while the subject moves a jointed manipulandum on a 2D plane to control a cursor so that
it hits randomly placed targets. (b) Decoding: true target trajectory (dashed (red): dark to
light) and reconstruction using the Kalman filter (solid (blue): dark to light).
end, we propose a Kalman filtering method that provides a rigorous and well understood
framework that addresses these issues. This approach provides a control-theoretic model
for the encoding of hand movement in motor cortex and for inferring, or decoding, this
movement from the firing rates of a population of cells.
Simultaneous recordings are acquired from an array consisting of microelectrodes [6]
implanted in the arm area of primary motor cortex (MI) of a Macaque monkey; recordings
from this area have been used previously to control devices [5, 9, 10, 11, 14]. The monkey
views a computer monitor while gripping a two-link manipulandum that controls the 2D
motion of a cursor on the monitor (Figure 1a). We use the experimental paradigm of [9], in
which a target dot appears in a random location on the monitor and the task requires moving
a feedback dot with the manipulandum so that it hits the target. When the target is hit, it
jumps to a new random location. The trajectory of the hand and the neural activity of
cells are recorded simultaneously. We compute the position, velocity, and acceleration of
the hand along with the mean firing rate for each of the cells within non-overlapping
time bins. In contrast to related work [8, 15], the motions of the monkey in this task are
quite rapid and more ?natural? in that the actual trajectory of the motion is unconstrained.
The reconstruction of hand trajectory from the mean firing rates can be viewed probabilistically as a problem of inferring behavior from noisy measurements. In [15] we proposed
a Kalman filter framework [3] for modeling the relationship between firing rates in motor
cortex and the position and velocity of the subject?s hand. This work focused on off-line
reconstruction using constrained motions of the hand [8]. Here we consider new data from
the on-line environmental setup [9] which is more natural, varied, and contains rapid motions. With this data we show that, in contrast to our previous results, a model of hand
acceleration (in addition to position and velocity) is important for accurate reconstruction.
In the Kalman framework, the hand movement (position, velocity and acceleration) is modeled as the system state and the neural firing rate is modeled as the observation (measurement). The approach specifies an explicit generative model that assumes the observation
(firing rate in ) is a linear function of the state (hand kinematics) plus Gaussian noise .
Similarly, the hand state at time
is assumed to be a linear function of the hand state at the
previous time instant plus Gaussian noise. The Kalman filter approach provides a recursive,
on-line, estimate of hand kinematics from the firing rate in non-overlapping time bins. The
This is a crude assumption but the firing rates can be square-root transformed [7] making them
more Gaussian and the mean firing rate can be subtracted to achieve zero-mean data.
results of reconstructing hand trajectories from pre-recorded neural firing rates are compared with those obtained using more traditional fixed linear filtering techniques [9, 12]
using overlapping windows. The results indicate that the Kalman filter decoding is
more accurate than that of the fixed linear filter.
1.1 Related Work
Georgopoulos and colleagues [4] showed that hand movement direction may be encoded
by the neural ensemble in the arm area of motor cortex (MI). This early work has resulted
in a number of successful algorithms for decoding neural activity in MI to perform offline reconstruction or on-line control of cursors or robotic arms. Roughly, the primary
methods for decoding MI activity include the population vector algorithm [4, 5, 7, 11],
linear filtering [9, 12], artificial neural networks [14], and probabilistic methods [2, 10, 15].
This population vector approach is the oldest method and it has been used for the real-time
neural control of 3D cursor movement [11]. This work has focused primarily on ?center
out? motions to a discrete set of radial targets (in 2D or 3D) rather than natural, continuous,
motion that we address here.
Linear filtering [8, 12] is a simple statistical method that is effective for real-time neural
control of a 2D cursor [9]. This approach requires the use of data over a long time win
dow (typically to ). The fixed linear filter, like population vectors and neural
networks [14] lack both a clear probabilistic model and a model of the temporal hand kinematics. Additionally, they provide no estimate of uncertainty and hence may be difficult to
extend to the analysis of more complex temporal movement patterns.
We argue that what is needed is a probabilistically grounded method that uses data in small
time windows (e.g. or less) and integrates that information over time in a recursive fashion. The C ONDENSATION algorithm has been recently introduced as a Bayesian
decoding scheme [2], which provides a probabilistic framework for causal estimation and
is shown superior to the performance of linear filtering when sufficient data is available
(e.g. using firing rates for several hundred cells). Note that the C ONDENSATION method is
more general than the Kalman filter proposed here in that it does not assume linear models
and Gaussian noise. While this may be important for neural decoding as suggested in [2],
current technology makes the method impractical for real-time control.
For real-time neural control we exploit the Kalman filter [3, 13] which has been widely
used for estimation problems ranging from target tracking to vehicle control. Here we
apply this well understood theory to the problem of decoding hand kinematics from neural
activity in motor cortex. This builds on the work that uses recursive Bayesian filters to
estimate the position of a rat from the firing activity of hippocampal place cells [1, 16]. In
contrast to the linear filter or population vector methods, this approach provides a measure
of confidence in the resulting estimates. This can be extremely important when the output
of the decoding method is to be used for later stages of analysis.
2 Methods
Decoding involves estimating the state of the hand at the current instant in time; i.e.
x
representing -position, -position, -velocity, -velocity, acceleration, and -acceleration at time
"!
where !
# in our experiments.
The Kalman filter [3, 13] model assumes the state is linearly related to the observations
z %$'&)( which here represents a *,+
vector containing the firing rates at time
for *
observed neurons within . In our experiments, * cells. We briefly review the
Kalman filter algorithm below; for details the reader is referred to [3, 13].
Encoding: We define a generative model of neural firing as
z
x q
(1)
" , is the number of time steps in the trial, and
$ & (
is a
where
matrix that linearly relates the hand state to the neural firing. We assume the noise in the
observations is zero mean and normally distributed, i.e. q
" % $ &)( ( .
The states are assumed to propagate in time according to the system model
x
x w
(2)
where $ & is the coefficient matrix and the noise term w
"
$
& . This states that the hand kinematics (position, velocity, and acceleration) at time
is linearly related to the state at time . Once again we assume these estimates are
normally distributed.
In practice, might change with time step , however, here we make the
common simplifying assumption they are constant. Thus we can estimate the Kalman filter
model from training data using least squares estimation:
!#"
!
$ '& &
$ &*&
'
&
&
)
x &'& )
argmin
x ( x
z
argmin
A
H
%
%
&*& &'&
)
where is the + norm. Given and it is then simple to estimate the noise covariance
matrices and ; details are given in [15].
Decoding: At each
" time step the algorithm has two steps: 1) prediction of the a priori
state estimate x, ; and 2) updating this estimate with new measurement data to produce an
a posteriori state estimate x, . In particular, these steps are:
I. Discrete Kalman filter time update equations:
At each time
, we obtain the" a priori estimate from the previous time
its error covariance matrix, - :
"
x, "
x,
"
- .- "
"
, then compute
(3)
(4)
II. Measurement update equations:
"
Using the estimate x, and firing rate z , we update the estimate using the measurement
and compute the posterior error covariance matrix:
x, /x,
-
"
"
0 z x,
"
21 0 3 -
(5)
(6)
where - represents the state error covariance after taking into account the neural data and
0 is the Kalman gain matrix given by
0
-
"
"
-
#
"
(7)
This 0 produces a state estimate that minimizes the mean squared error of the reconstruction (see [3] for details). Note that is the measurement error matrix and, depending on
the reliability of the data, the gain term, 0 , automatically adjusts the contribution of the
new measurement to the state estimate.
)
Method
Correlation Coefficient
MSE ( )
Kalman (0 lag)
(0.768, 0.912)
7.09
Kalman (70 lag)
(0.785, 0.932)
7.07
Kalman (140 lag)
(0.815, 0.929)
6.28
Kalman (210 lag)
(0.808, 0.891)
6.87
Kalman (no acceleration)
(0.817, 0.914)
6.60
Linear filter
(0.756, 0.915)
8.30
Table 1: Reconstruction results for the fixed linear and recursive Kalman filter. The table
also shows how the Kalman filter results vary with lag times (see text).
3 Experimental Results
To be practical, we must be able to train the model (i.e. estimate , , , ) using a
small amount of data. Experimentally we found that approximately 3.5 minutes of training
data suffices for accurate reconstruction (this is similar to the result for fixed linear filters
reported in [9]). As described in the introduction, the task involves moving a manipulan
dum freely on a + tablet (with a + workspace) to hit randomly
placed targets on the screen. We gather the mean firing rates and actual hand trajectories
for the training data and then learn the models via least squares (the computation time is
negligible). We then test the accuracy of the method by reconstructing test trajectories offline using recorded neural data not present in the training set. The results reported here use
approximately 1 minute of test data.
Optimal Lag: The physical relationship between neural firing and arm movement means
there exists a time lag between them [7, 8]. The introduction of a time lag results in the
measurements, z , at time
, being taken from some previous (or future) instant in time
" for some integer . In the interest of simplicity, we consider a single optimal time
lag for all the cells though evidence suggests that individual time lags may provide better
results [15].
Using time lags of 0, 70, 140, 210 we train the Kalman filter and perform reconstruction
(see Table 1). We report the accuracy of the reconstructions with a variety of error measures
used in the literature including the correlation coefficient ( ) and the mean squared error
(MSE) between the reconstructed and true trajectories. From Table 1 we see that optimal
lag is around two time steps (or 140 ); this lag will be used in the remainder of the
experiments and is similar to our previous findings [15] which suggested that the optimal
lag was between 50-100 .
Decoding: At the beginning of the test trial we let the predicted initial condition equal the
real initial condition. Then the update equations in Section 2 are applied. Some examples of
the reconstructed trajectory are shown in Figure 2 while Figure 3 shows the reconstruction
of each component of the state variable (position, velocity and acceleration in and ).
From Figure 3 and Table 1 we note that the reconstruction in is more accurate than in
the direction (the same is true for the fixed linear filter described below); this requires
further investigation. Note also that the ground truth velocity and acceleration curves are
computed from the position data with simple differencing. As a result these plots are quite
noisy making an evaluation of the reconstruction difficult.
20
20
20
18
18
18
16
16
16
14
14
14
12
12
12
10
10
10
8
8
8
6
6
6
4
4
4
2
2
2
4
6
8
10
12
14
16
18
20
22
2
2
4
6
8
10
12
14
16
18
20
22
2
4
6
8
10
12
14
16
18
20
22
Figure 2: Reconstructed trajectories (portions of 1min test data ? each plot shows 50 time
instants (3.5 )): true target trajectory (dashed (red)) and reconstruction using the Kalman
filter (solid (blue)).
3.1 Comparison with linear filtering
Fixed linear filters reconstruct hand position as a linear combination of the firing rates over
some fixed time period [4, 9, 12]; that is,
$
$
%
"
where is the -position (or, equivalently, the -position) at time
!
!
, , where
is the number of time steps in a trial, is the constant
offset, " is the firing rate of neuron at time
" , and are the filter coefficients. The
coefficients can be learned from training data using a simple least squares technique. In
our experiments here we take which means that the hand position is determined
from firing data over . This is exactly the method described in [9] which provides a fair
comparison for the Kalman filter; for details see [12, 15]. Note that since the linear filter
uses data over a long time window, it does not benefit from the use of time-lagged data.
Note also that it does not explicitly reconstruct velocity or acceleration.
The linear filter reconstruction of position is shown in Figure 4. Compared with Figure 3,
we see that the results are visually similar. Table 1, however, shows that the Kalman filter
gives a more accurate reconstruction than the linear filter (higher correlation coefficient and
lower mean-squared error). While fixed linear filtering is extremely simple, it lacks many
of the desirable properties of the Kalman filter.
Analysis: In our previous work [15], the experimental paradigm involved carefully designed hand motions that were slow and smooth. In that case we showed that acceleration
was redundant and could be removed from the state equation. The data used here is more
?natural?, varied, and rapid and we find that modeling acceleration improves the prediction
of the system state and the accuracy of the reconstruction; Table 1 shows the decrease in
accuracy with only position and velocity in the system state (with 140ms lag).
4 Conclusions
We have described a discrete linear Kalman filter that is appropriate for the neural control
of 2D cursor motion. The model can be easily learned using a few minutes of training data
and provides real-time estimates of hand position every given the firing rates of 42
x-position
y-position
20
10
15
5
10
0
5
5
10
15
20
5
x-velocity
10
15
20
15
20
15
20
y-velocity
2
2
1
0
0
1
2
2
5
10
15
20
5
x-acceleration
10
y-acceleration
2
1
1
0
0
1
1
2
5
10
15
20
5
time (second)
10
time (second)
Figure 3: Reconstruction of each component of the system state variable: true target motion
(dashed (red)) and reconstruction using the Kalman filter (solid (blue)). 20 from a 1min
test sequence are shown.
y-position
x-position
20
10
15
5
10
0
5
5
10
time (second)
15
20
5
10
15
20
time (second)
Figure 4: Reconstruction of position using the linear filter: true target trajectory (dashed
(red)) and reconstruction using the linear filter (solid (blue)).
cells in primary motor cortex. The estimated trajectories are more accurate than the fixed
linear filtering results being used currently.
The Kalman filter proposed here provides a rigorous probabilistic approach with a well
understood theory. By making its assumptions explicit and by providing an estimate of
uncertainty, the Kalman filter offers significant advantages over previous methods. The
method also estimates hand velocity and acceleration in addition to 2D position. In contrast
to previous experiments, we show, for the natural 2D motions in this task, that incorporating acceleration into the system and measurement models improves the accuracy of the
decoding. We also show that, consistent with previous studies, a time lag of
improves the accuracy.
Our future work will evaluate the performance of the Kalman filter for on-line neural control of cursor motion in the task described here. Additionally, we are exploring alternative
measurement noise models, non-linear system models, and non-linear particle filter decod-
ing methods. Finally, to get a complete picture of current methods, we are pursuing further
comparisons with population vector methods [7] and particle filtering techniques [2].
Acknowledgments. This work was supported in part by: the DARPA Brain Machine
Interface Program, NINDS Neural Prosthetics Program and Grant #NS25074, and the National Science Foundation (ITR Program award #0113679). We thank J. Dushanova, C.
Vargas, L. Lennox, and M. Fellows for their assistance.
References
[1] Brown, E., Frank, L., Tang, D., Quirk, M., and Wilson, M. (1998). A statistical paradigm for
neural spike train decoding applied to position prediction from ensemble firing patterns of rat
hippocampal place cells. J. of Neuroscience, 18(18):7411?7425.
[2] Gao, Y., Black, M. J., Bienenstock, E., Shoham, S., and Donoghue, J. P. (2002). Probabilistic
inference of hand motion from neural activity in motor cortex. Advances in Neural Information
Processing Systems 14, The MIT Press.
[3] Gelb, A., (Ed.) (1974). Applied Optimal Estimation. MIT Press.
[4] Georgopoulos, A., Schwartz, A., and Kettner, R. (1986). Neural population coding of movement direction. Science, 233:1416?1419.
[5] Helms Tillery, S., Taylor, D., Isaacs, R., Schwartz, A. (2000) Online control of a prosthetic
arm from motor cortical signals. Soc. for Neuroscience Abst., Vol. 26.
[6] Maynard, E., Nordhausen C., Normann, R. (1997). The Utah intracortical electrode array: A
recording structure for potential brain-computer interfaces. Electroencephalography and Clinical Neuophysiology 102, pp. 228?239.
[7] Moran, D. and Schwartz, B. (1999). Motor cortical representation of speed and direction during
reaching. J. of Neurophysiology, 82(5):2676?2692.
[8] Paninski, L., Fellows, M., Hatsopoulos, N., and Donoghue, J. P. (2001). Temporal tuning
properties for hand position and velocity in motor cortical neurons. submitted, J. of Neurophysiology.
[9] Serruya, M. D., Hatsopoulos, N. G., Paninski, L., Fellows, M. R., and Donoghue, J. P. (2002).
Brain-machine interface: Instant neural control of a movement signal. Nature, (416):141?142.
[10] Serruya. M., Hatsopoulos, N., Donoghue, J., (2000) Assignment of primate M1 cortical activity
to robot arm position with Bayesian reconstruction algorithm. Soc. for Neuro. Abst., Vol. 26.
[11] Taylor. D., Tillery, S., Schwartz, A. (2002). Direct cortical control of 3D neuroprosthetic
devices. Science, Jun. 7;296(5574):1829-32.
[12] Warland, D., Reinagel, P., and Meister, M. (1997). Decoding visual information from a population of retinal ganglion cells. J. of Neurophysiology, 78(5):2336?2350.
[13] Welch, G. and Bishop, G. (2001). An introduction to the Kalman filter. Technical Report TR
95-041, University of North Carolina at Chapel Hill, Chapel Hill,NC 27599-3175.
[14] Wessberg, J., Stambaugh, C., Kralik, J., Beck, P., Laubach, M., Chapin, J., Kim, J., Biggs, S.,
Srinivasan, M., and Nicolelis, M. (2000). Real-time prediction of hand trajectory by ensembles
of cortical neurons in primates. Nature, 408:361?365.
[15] Wu, W., Black, M. J., Gao, Y., Bienenstock, E., Serruya, M., and Donoghue, J. P., Inferring hand motion from multi-cell recordings in motor cortex using a Kalman filter, SAB?02Workshop on Motor Control in Humans and Robots: On the Interplay of Real Brains and
Artificial Devices, Aug. 10, 2002, Edinburgh, Scotland, pp. 66?73.
[16] Zhang, K., Ginzburg, I., McNaughton, B., Sejnowski, T., Interpreting neuronal population
activity by reconstruction: Unified framework with application to hippocampal place cells, J.
Neurophysiol. 79:1017?1044, 1998.
| 2178 |@word neurophysiology:3 trial:3 briefly:1 norm:1 propagate:1 carolina:1 simplifying:1 covariance:4 tr:1 solid:4 initial:2 contains:1 nordhausen:1 current:3 must:2 john:1 realistic:1 motor:13 plot:2 designed:1 update:4 generative:2 device:5 manipulandum:4 wessberg:1 plane:1 oldest:1 beginning:1 scotland:1 short:1 provides:9 location:2 zhang:1 mathematical:1 along:1 direct:3 acquired:1 rapid:3 roughly:1 behavior:1 multi:1 brain:4 automatically:1 actual:2 window:3 electroencephalography:1 estimating:1 chapin:1 what:1 argmin:2 minimizes:1 monkey:3 unified:1 finding:1 impractical:1 temporal:3 fellow:3 every:1 exactly:1 hit:5 schwartz:4 control:21 normally:2 grant:1 negligible:1 understood:3 encoding:3 firing:25 approximately:3 black:4 plus:2 might:1 suggests:1 elie:1 practical:1 acknowledgment:1 shaikhouni:2 recursive:4 practice:1 area:4 shoham:1 pre:1 radial:1 confidence:1 get:1 dam:1 demonstrated:1 center:1 focused:2 welch:1 simplicity:1 chapel:2 insight:2 adjusts:1 array:3 reinagel:1 enabled:1 population:9 mcnaughton:1 target:13 us:3 tablet:2 velocity:15 updating:1 observed:1 movement:12 removed:1 decrease:1 hatsopoulos:3 division:2 biggs:1 neurophysiol:1 easily:1 darpa:1 train:3 effective:1 sejnowski:1 artificial:2 quite:2 encoded:1 widely:1 lag:16 reconstruct:2 noisy:2 online:1 interplay:1 advantage:2 sequence:2 abst:2 reconstruction:24 propose:1 remainder:1 achieve:1 decod:1 tillery:2 electrode:3 produce:3 depending:1 develop:2 quirk:1 aug:1 soc:2 c:1 involves:2 indicate:2 predicted:1 direction:4 filter:42 human:1 bin:3 require:1 suffices:1 investigation:1 exploring:1 stambaugh:1 around:1 ground:1 visually:1 cfm:2 vary:1 early:1 estimation:4 integrates:1 currently:1 mit:2 gaussian:4 rather:1 reaching:1 sab:1 wilson:1 probabilistically:2 focus:1 prosthetics:1 contrast:5 rigorous:2 mijail:1 kim:1 posteriori:1 inference:1 typically:1 bienenstock:3 transformed:1 issue:1 priori:2 constrained:1 equal:1 once:1 biology:1 represents:2 future:2 report:2 primarily:1 few:1 randomly:3 microelectrodes:1 simultaneously:1 resulted:1 national:1 individual:1 beck:1 consisting:1 interest:1 evaluation:1 light:2 accurate:8 taylor:2 causal:1 minimal:2 modeling:2 assignment:1 hundred:1 delay:1 successful:1 reported:3 providence:1 workspace:1 probabilistic:7 off:2 decoding:20 again:1 squared:3 recorded:5 containing:1 external:1 account:1 potential:1 intracortical:1 retinal:1 coding:3 north:1 coefficient:6 explicitly:3 vehicle:1 root:1 view:1 later:1 red:4 portion:1 contribution:1 square:4 accuracy:6 ensemble:3 bayesian:3 trajectory:16 submitted:1 simultaneous:1 ed:1 colleague:1 pp:2 involved:1 isaac:1 mi:4 gain:2 improves:3 carefully:1 appears:1 higher:1 though:1 stage:1 correlation:3 hand:35 dow:1 overlapping:3 lack:2 maynard:1 utah:1 brown:9 true:6 hence:1 assistance:1 during:1 rat:2 m:2 hippocampal:3 hill:2 theoretic:2 complete:1 motion:18 interface:3 interpreting:1 ranging:1 recently:1 superior:1 common:1 spiking:2 physical:1 extend:1 m1:1 measurement:11 significant:1 tuning:1 unconstrained:1 mathematics:1 similarly:1 particle:2 dot:2 reliability:1 moving:2 robot:2 cortex:9 posterior:1 recent:1 showed:2 freely:1 paradigm:3 period:1 redundant:1 signal:2 dashed:4 relates:1 ii:1 desirable:1 sound:1 smooth:1 ing:1 technical:1 clinical:1 offer:1 long:2 dept:2 award:1 feasibility:1 prediction:4 neuro:1 implanted:3 grounded:1 serruya:5 achieved:1 cell:14 addition:2 subject:3 recording:4 integer:1 variety:2 itr:1 donoghue:7 clear:1 amount:2 dark:2 specifies:1 neuroscience:3 estimated:1 blue:4 discrete:3 vol:2 srinivasan:1 monitor:5 uncertainty:3 place:3 reader:1 wu:2 pursuing:1 jointed:1 ninds:1 gripping:1 display:1 activity:11 georgopoulos:2 ri:1 prosthetic:2 speed:1 extremely:2 min:2 vargas:1 according:1 combination:1 reconstructing:3 making:3 primate:2 ginzburg:1 taken:1 equation:4 previously:2 ondensation:2 kinematics:5 needed:1 end:1 meister:1 available:1 apply:1 limb:1 appropriate:1 subtracted:1 alternative:1 assumes:2 include:1 instant:5 medicine:1 exploit:1 warland:1 build:1 move:2 spike:1 primary:4 traditional:1 win:1 link:1 thank:1 argue:2 kalman:35 dum:1 modeled:2 relationship:3 providing:1 laubach:1 differencing:1 setup:1 difficult:2 equivalently:1 nc:1 frank:1 lagged:1 kralik:1 perform:2 neuron:5 observation:4 gelb:1 varied:2 introduced:1 learned:2 macaque:1 address:2 able:1 suggested:2 below:2 pattern:2 program:3 including:1 natural:5 nicolelis:1 arm:7 representing:2 scheme:1 technology:1 picture:1 jun:1 normann:1 text:1 review:1 literature:1 ammar:1 filtering:10 foundation:2 gather:1 sufficient:1 consistent:1 placed:3 supported:1 offline:2 taking:1 distributed:2 benefit:1 feedback:1 curve:1 cortical:6 neuroprosthetic:1 edinburgh:1 collection:1 jump:1 reconstructed:3 robotic:1 assumed:2 continuous:2 helm:1 table:7 additionally:2 nature:3 learn:1 kettner:1 mse:2 complex:1 linearly:3 noise:8 fair:1 neuronal:1 referred:1 screen:1 fashion:1 slow:1 position:29 inferring:3 explicit:2 crude:1 tang:1 minute:3 bishop:1 moran:1 offset:1 evidence:1 exists:1 incorporating:1 workshop:1 cursor:11 paninski:2 gao:4 ganglion:1 visual:1 tracking:1 truth:1 environmental:1 viewed:1 acceleration:16 change:1 experimentally:1 determined:1 experimental:3 evaluate:1 |
1,294 | 2,179 | Mismatch String Kernels for SVM Protein
Classification
Christina Leslie
Department of Computer Science
Columbia University
[email protected]
Eleazar Eskin
Department of Computer Science
Columbia University
[email protected]
Jason Weston
Max-Planck Institute
Tuebingen, Germany
[email protected]
William Stafford Noble
Department of Genome Sciences
University of Washington
[email protected]
Abstract
We introduce a class of string kernels, called mismatch kernels, for use
with support vector machines (SVMs) in a discriminative approach to
the protein classification problem. These kernels measure sequence similarity based on shared occurrences of -length subsequences, counted
with up to mismatches, and do not rely on any generative model for
the positive training sequences. We compute the kernels efficiently using
a mismatch tree data structure and report experiments on a benchmark
SCOP dataset, where we show that the mismatch kernel used with an
SVM classifier performs as well as the Fisher kernel, the most successful method for remote homology detection, while achieving considerable
computational savings.
1 Introduction
A fundamental problem in computational biology is the classification of proteins into functional and structural classes based on homology (evolutionary similarity) of protein sequence data. Known methods for protein classification and homology detection include
pairwise sequence alignment [1, 2, 3], profiles for protein families [4], consensus patterns
using motifs [5, 6] and profile hidden Markov models [7, 8, 9]. We are most interested
in discriminative methods, where protein sequences are seen as a set of labeled examples
? positive if they are in the protein family or superfamily and negative otherwise ? and
we train a classifier to distinguish between the two classes. We focus on the more difficult
problem of remote homology detection, where we want our classifier to detect (as positives)
test sequences that are only remotely related to the positive training sequences.
One of the most successful discriminative techniques for protein classification ? and the
best performing method for remote homology detection ? is the Fisher-SVM [10, 11] approach of Jaakkola et al. In this method, one first builds a profile hidden Markov model
Formerly William Noble Grundy: see http://www.cs.columbia.edu/?noble/name-change.html
(HMM) for the positive training sequences, defining a log likelihood function
for any protein sequence . If
is the maximum likelihood estimate for the model parameters, then the gradient vector
assigns to each (positive or negative)
training sequence an explicit vector of features called Fisher scores. This feature mapping defines a kernel function, called the Fisher kernel, that can then be used to train a
support vector machine (SVM) [12, 13] classifier. One of the strengths of the Fisher-SVM
approach is that it combines the rich biological information encoded in a hidden Markov
model with the discriminative power of the SVM algorithm. However, one generally needs
a lot of data or sophisticated priors to train the hidden Markov model, and because calculating the Fisher scores requires computing forward and backward probabilities from the
Baum-Welch algorithm (quadratic in sequence length for profile HMMs), in practice it is
very expensive to compute the kernel matrix.
In this paper, we present a new string kernel,"called
the mismatch kernel, for use with an
!
SVM for remote homology detection. The -mismatch kernel is based on a feature
map to a vector space indexed by all possible subsequences of amino acids of a fixed
length ; each instance of a fixed -length subsequence in an input sequence contributes
to all feature coordinates differing from it by at most mismatches. Thus, the mismatch
kernel adds the biologically important idea of mismatching to the computationally simpler
spectrum kernel presented in [14]. In the current work, we also describe how"to
compute
!
the new kernel efficiently using a mismatch tree data structure; for values of useful
in this application, the kernel is fast enough to use on real datasets and is considerably
less expensive than the Fisher kernel. We report results from a benchmark dataset on the
SCOP database [15] assembled by Jaakkola et al. [10] and show that the mismatch kernel
used with an SVM classifier achieves performance equal to the Fisher-SVM method while
outperforming all other methods tested. Finally, we note that the mismatch kernel does
not depend on any generative model and could potentially be used in other sequence-based
classification problems.
2 Spectrum and Mismatch String Kernels
The basis for our approach to protein classification is to represent protein sequences as
vectors in a high-dimensional feature space via a string-based feature map. We then train
a support vector machine (SVM), a large-margin linear classifier, on the feature vectors
representing our training sequences. Since SVMs are a kernel-based learning algorithm,
we do not calculate the feature vectors explicitly but instead compute their pairwise inner
products using a mismatch string kernel, which we define in this section.
2.1 Feature Maps for Strings
"!
The # -mismatch kernel is based on a feature map from the space of all finite sequences
from an alphabet $ of size $%&(' to the ') -dimensional vector space indexed by the set
of -length subsequences (? -mers?) from $ . (For protein sequences, $ is the alphabet of
amino
acids, '*&,+- .) For a fixed -mer ./&,01204365758590 , with each 0;: a character in $ ,
"!
)
the -neighborhood generated by . is the set of all -length sequences < from $ that
E. .
differ from . by at most mismatches. We denote this set by =?>
)A@ BDC
We define our feature map FG>
)A@ BDC
FH>
as follows: if . is a -mer, then
)A@ BDC
(1)
E.&IKJMLNE.9 L;OPQ
where JRLE.S&UT if < belongs to =V>
-mer
E. , and JLW.X&Y- otherwise. Thus, a
)A@ BDC
contributes weight to all the coordinates in its mismatch neighborhood.
For a sequence
Z
of any length, we extend the map additively by summing the feature
vectors for all the -mers in Z :
FH>
KZ"
) @ B*C
&
FH>
)
-mers in
W.
)A@ BDC
Note that the < -coordinate of FG>
-mer <
KZR is just a count of all instances of the
"!
)A@ BDC
occurring with up to mismatches in Z . The # -mismatch kernel >
is the inner
) @ B*C
product in feature space of feature vectors:
>
!
KZ
)A@ BDC
KFH>
&
) @ B*C
!
KZ"
FH>
)A@ BDC
25
For
&/
, we retrieve the -spectrum kernel defined in [14].
2.2 Fisher Scores and the Spectrum Kernel
While we define the spectrum and mismatch feature maps without any reference to a generative model for the positive class of sequences, there is some similarity
between the
-spectrum feature map and the Fisher scores associated to an order
T Markov chain
model. More precisely, suppose the generative model for the positive training sequences is
given by
KZ
&
for a string ZV&
1
58575
1 3658575
1
E
&
7
!
!
!
57575
1
!
585759
1
")
&
1
"58575 E
1
)
)
, with parameters
)
!
57585
1
)
1
")
!
575859
1
1
!"" Q #
&
for characters 1 58575
1 in alphabet $ . Denote by
the maximum likelihood es)
timate for
on the positive training set. To calculate the Fisher scores for this model,
$&% ' ( ( ( ' Q#
we follow [10] and define independent variables
@
!"" Q!# & )
$ * % '
' Q!# satisfying
@ "" !Q #
&/
"" !Q # , ,
*
*
Q #
@ ""
.
WZ
@ "" Q#
&
.
&
&
T
. Then the Fisher scores are given by
/
T
!"" Q!#
( ( (
$+*
"" Q!#
"" Q!#
.
"" Q!#
Q!#
@ ""
0
@
!
"" Q#
1
1
.
32
"" !Q #
where .
is the number of instances of the -mer 1 57585
1 in Z , and .
)
4
3!"" Q#
"" Q!#
is the number of instances of the
T -mer M157585
1 . Thus the Fisher score captures the
)
degree to which the -mer 158575!
16 is over- or under-represented relative to the positive
)5
model. For the -spectrum kernel, the corresponding feature coordinate looks similar but
KZ" &7.
5
simply uses the unweighted count: J
Q!#
Q!#
!""
""
3 Efficient Computation of the Mismatch Kernel
Unlike the Fisher vectors used in [10], our feature vectors are sparse vectors in a very high
dimensional feature space. Thus, instead of calculating and storing the feature vectors, we
directly and efficiently compute the kernel matrix for use with an SVM classifier.
3.1 Mismatch Tree Data Structure
We use a mismatch tree data structure (similar to a trie or suffix tree [16, 17]) to represent
the feature space (the set of all -mers) and perform a lexical traversal of all -mers occurring in the sample dataset match with up to of mismatches; the entire kernel matrix
KZA:
!
Z
,
!
&
for the sample of
58575
T
sequences is computed in one traversal of
the tree.
!
A # -mismatch tree is a rooted tree of depth where each internal node has $ "& '
branches and each branch is labeled with a symbol from $ . A leaf node represents a fixed
-mer in our feature space ? obtained by concatenating the branch symbols along the path
from root to leaf ? and an internal node represents the prefix for those -mer features which
are its descendants in the tree. We use a depth-first search of this tree to store, at each node
that we visit, a set of pointers to all instances of the current prefix pattern that occur with
mismatches in the sample data. Thus at each node of depth , we maintain pointers to
all substrings from the sample data set whose -length prefixes are within mismatches
from the -length prefix represented by the path down from the root. Note that the set of
valid substrings at a node is a subset of the set of valid substrings of its parent. When we
encounter a node with an empty list of pointers (no valid occurrences of the current prefix),
we do not need to search below it in the tree. When we reach a leaf node, we sum the
contributions of all instances occurring in each source sequence to obtain feature! values
corresponding to the current -mer, and we update the kernel matrix entry WZ Z for
each pair of source sequences Z
and Z having non-zero feature values.
0
A
V
L
A
L
K
A
V
A
0
A
V
L
A
L
K
A
V
A
0
A
V
L
A
L
K
A
V
(a)
0
V
L
A
L
K
A
V
L
0
L
A
L
K
A
V
L
L
(b)
0
V
L
A
L
K
A
V
1
L
A
L
K
A
V
L
1
A
L
K
A
V
L
L
0
V
L
A
L
K
A
V
L
0
V
L
A
L
K
A
V
0
L
A
L
K
A
V
L
L
1
L
A
L
K
A
V
L
1
A
L
K
A
V
L
L
0
V
L
A
L
K
A
V
L
0
L
A
L
K
A
V
L
L
L
1
L
A
L
K
A
V
(c)
1
A
L
K
A
V
L
Figure 1: An
-mismatch tree for a sequence AVLALKAVLL, showing valid instances at each
node down a path: (a) at the root node; (b) after expanding the path ; and (c) after expanding the
path . The number of mismatches for each instance is also indicated.
3.2 Efficiency of the Kernel Computation
Since we compute the kernel in one depth-first traversal, we do not actually need to store
the entire mismatch tree but instead compute the kernel using a recursive function, which
makes more efficient use of memory and allows kernel computations for large datasets.
,
B
:
E'
"!
The number
of -mers within
TA
:
mismatches of any given fixed -mer is
&
!
N
'K&
B
'WB
. Thus the effective number of -mer instances that we
B 'WB , where =
need to traverse grows as E=
is the total length of the sample data. At
a leaf node, if exactly input sequences contain valid instances of the current -mer, one
3
performs updates to the kernel matrix. For sequences each of length . (total length
=
&.!
), the worst case for the kernel computation occurs when the feature vectors
are all equal and have the
maximal
number of non-zero entries, giving worst case overall
"!
!
3
3
running time "
.#
'W &
"
. B ' B . For the application we discuss here,
small values of are most useful, and the kernel calculations are quite inexpensive.
When mismatch kernels are used in combination with SVMs, the learned classifier $WZ"6&
,
!
(where Z : are the training sequences that map to
FH>
WZ"3
)A@ BDC
support vectors, : are labels, and . : are weights) can be implemented by pre-computing
and storing per -mer scores. Then the prediction $WZ" can be calculated in linear time by
look-up of -mer scores. In practice, one usually wants to use a normalized feature map,
so one would also need to compute the norm of the vector F>
WZ" , with complexity
)A@ BDC
.
B 'WB for a sequence of length . . Simple
9TA normalization schemes, like dividing
by sequence length, can also be used.
:
:W.:
1
FH>
KZ8:K
)A@ BDC
4 Experiments: Remote Protein Homology Detection
We test the mismatch kernel with an SVM classifier on the SCOP [15] (version 1.37)
datasets designed by Jaakkola et al. [10] for the remote homology detection problem. In
these experiments, remote homology is simulated by holding out all members of a target
SCOP family from a given superfamily. Positive training examples are chosen from the
remaining families in the same superfamily, and negative test and training examples are
chosen from disjoint sets of folds outside the target family?s fold. The held-out family
members serve as positive test examples. In order to train HMMs, Jaakkola et al. used the
SAM-T98 algorithm to pull in domain homologs from the non-redundant protein database
and added these sequences as positive examples in the experiments. Details of the datasets
are available at www.soe.ucsc.edu/research/compbio/discriminative.
Because the test sets are
designed ! for remote! homology detection, we use small val"!
ues of . We tested # & ! TA and TA , where we normalized the kernel via
>
Norm
) @ B*C
KZ
!
&
>
#>
)A@ BDC
)A@ D
B C
!
Z
W
Z"
WZ
>
) @ B*C
M!
"!
5
We found that
&
!
TA
gave
slightly
better performance,
though results were similar for the two choices. (Data
"!
!
for & T not shown.) We use a publicly available SVM implementation
(www.cs.columbia.edu/compbio/svm) of the soft margin optimization algorithm described
in [10]. For comparison, we include results from three other methods. These include the
original experimental results from Jaakkola et al. for two methods: the SAM-T98 iterative
HMM, and the Fisher-SVM method. We also test PSI-BLAST [3], an alignment-based
method widely used in the biological community, on the same data using the methodology
described in [14].
Figure 2 illustrates the mismatch-SVM method?s performance relative to three existing
homology detection methods as measured by ROC scores. The figure includes results for
SCOP families, and each series corresponds to one homology detection method.
all
Qualitatively, the curves for Fisher-SVM and mismatch-SVM are quite similar. When
we compare the overall performance of two methods using a two-tailed signed rank test
[18, 19] based on ROC scores over the 33 families with a -value threshold of -M5 - and
including a Bonferroni adjustment to account for multiple comparisons, we find only the
following significant differences: Fisher-SVM and mismatch-SVM perform better than
SAM-T98 (with p-values 1.3e-02 and 2.7e-02, respectively); and these three methods all
perform significantly better than PSI-BLAST in this experiment.
!
Figure 3 shows a family-by-family comparison of performance of the TA -mismatchSVM and Fisher-SVM using ROC scores in plot (A) and ROC-50 scores in plot (B). 1 In
both plots, the points fall approximately evenly above and below the diagonal, indicating
little difference in performance between the two methods. Figure 4 shows the improvement
provided by including mismatches in the SVM kernel. The figures plot ROC scores (plot
1
The ROC-50 score is the area under the graph of the number of true positives as a function of
false positives, up to the first 50 false positives, scaled so that both axes range from 0 to 1. This
score is sometimes preferred in the computational biology community, motivated by the idea that a
biologist might be willing to sift through about 50 false positives.
35
Number of families
30
25
20
15
10
(5,1)-Mismatch-SVM ROC
Fisher-SVM ROC
SAM-T98
PSI-BLAST
5
0
0.5
0.55
0.6
0.65
0.7
0.75
0.8
0.85
0.9
0.95
1
ROC
Figure 2: Comparison of four homology detection methods. The graph plots the total number of
families for which a given method exceeds an ROC score threshold.
, & T
(A)) and ROC-50 scores (plot (B)) for two string kernel SVM methods: using &
mismatch kernel, and using &
(no mismatch) spectrum kernel, the best-performing
choice with
&
- . Almost all of the families perform better with mismatching than
without, showing that mismatching gives significantly better generalization performance.
5 Discussion
We have presented a class of string kernels that measure sequence similarity without requiring alignment or depending upon a generative model, and we have given an efficient
method for computing these kernels. For the remote homology detection problem, our discriminative approach ? combining support vector machines with the mismatch kernel ?
performs as well in the SCOP experiments as the most successful known method.
A practical protein classification system would involve fast multi-class prediction ? potentially involving thousands of binary classifiers ? on massive test sets. In such applications,
computational efficiency of the kernel function becomes an important issue. Chris Watkins
[20] and David Haussler [21] have recently defined a set of kernel functions over strings,
and one of these string kernels has been implemented for a text classification problem [22].
However, the cost of computing each kernel entry is . 3 in the length of the input sequences. Similarly, the Fisher kernel of Jaakkola
et al. requires quadratic-time computation
!
for each Fisher vector calculated. The # -mismatch kernel is relatively inexpensive to
compute for values of
that are practical in applications, allows computation of multiple kernel values in one pass, and significantly improves performance over the previously
presented (mismatch-free) spectrum kernel.
Many family-based remote homogy detection algorithms incorporate a method for selecting probable domain homologs from unannotated protein sequence databases for additional
training data. In these experiments, we used the domain homologs that were identified by
SAM-T98 (an iterative HMM-based algorithm) as part of the Fisher-SVM method and included in the datasets; these homologs may be more useful to the Fisher kernel than to the
mismatch kernel. We plan to extend our method by investigating semi-supervised techniques for selecting unannotated sequences for use with the mismatch-SVM.
1
1
0.95
0.8
Fisher-SVM ROC50
Fisher-SVM ROC
0.9
0.85
0.8
0.75
0.7
0.65
0.6
0.6
0.4
0.2
0.55
0.5
0.5
0
0.55
0.6
0.65 0.7 0.75 0.8 0.85
(5,1)-Mismatch-SVM ROC
0.9
0.95
1
0
0.2
(A)
0.4
0.6
(5,1)-Mismatch-SVM ROC50
0.8
1
(B)
Figure 3: Family-by-family comparison of -mismatch-SVM with Fisher-SVM. The coordinates of each point in the plot are the ROC scores (plot (A)) or ROC-50 scores (plot (B)) for one
SCOP family, obtained using the mismatch-SVM with
,
(x-axis) and Fisher-SVM
.
(y-axis). The dotted line is
1
1
0.95
k=3 Spectrum-SVM ROC50
k=3 Spectrum-SVM ROC
0.9
0.85
0.8
0.75
0.7
0.65
0.6
0.8
0.6
0.4
0.2
0.55
0.5
0.5
0
0.55
0.6
0.65
0.7
0.75
0.8
0.85
0.9
0.95
1
0
0.2
(5,1)-Mismatch-SVM ROC
0.4
0.6
0.8
1
(5,1)-Mismatch-SVM ROC50
(A)
(B)
Figure 4: Family-by-family comparison of -mismatch-SVM with spectrum-SVM. The coordinates of each point in the plot are the ROC scores (plot (A)) or ROC-50 scores (plot (B)) for one
SCOP family, obtained using the mismatch-SVM with
,
(x-axis) and spectrum-SVM
with
(y-axis). The dotted line is
.
Many interesting variations on the mismatch kernel can be explored using the framework
presented here. For example, explicit -mer feature selection can be implemented during calculation of the kernel matrix, based on a criterion enforced at each leaf or internal
node. Potentially, a good feature selection criterion could improve performance in certain
applications while decreasing kernel computation time. In biological applications, it is
also natural to consider weighting each -mer instance contribution to a feature coordinate
by evolutionary substitution probabilities. Finally, one could use linear combinations of
kernels #>
to capture similarity of different length -mers. We believe that further
) @ B EC
experimentation with mismatch string kernels could be fruitful for remote protein homology detection and other biological sequence classification problems.
Acknowledgments
CL is partially supported by NIH grant LM07276-02. WSN is supported by NSF grants
DBI-0078523 and ISI-0093302. We thank Nir Friedman for pointing out the connection
with Fisher scores for Markov chain models.
References
[1] M. S. Waterman, J. Joyce, and M. Eggert. Computer alignment of sequences, chapter Phylogenetic Analysis of DNA Sequences. Oxford, 1991.
[2] S. F. Altschul, W. Gish, W. Miller, E. W. Myers, and D. J. Lipman. A basic local alignment
search tool. Journal of Molecular Biology, 215:403?410, 1990.
[3] S. F. Altschul, T. L. Madden, A. A. Schaffer, J. Zhang, Z. Zhang, W. Miller, and D. J. Lipman. Gapped BLAST and PSI-BLAST: A new generation of protein database search programs.
Nucleic Acids Research, 25:3389?3402, 1997.
[4] Michael Gribskov, Andrew D. McLachlan, and David Eisenberg. Profile analysis: Detection of
distantly related proteins. PNAS, pages 4355?4358, 1987.
[5] A. Bairoch. The PROSITE database, its status in 1995. Nucleic Acids Research, 24:189?196,
1995.
[6] T. K. Attwood, M. E. Beck, D. R. Flower, P. Scordis, and J. N Selley. The PRINTS protein
fingerprint database in its fifth year. Nucleic Acids Research, 26(1):304?308, 1998.
[7] A. Krogh, M. Brown, I. Mian, K. Sjolander, and D. Haussler. Hidden markov models in computational biology: Applications to protein modeling. Journal of Molecular Biology, 235:1501?
1531, 1994.
[8] S. R. Eddy. Multiple alignment using hidden markov models. In Proceedings of the Third
International Conference on Intelligent Systems for Molecular Biology, pages 114?120. AAAI
Press, 1995.
[9] P. Baldi, Y. Chauvin, T. Hunkapiller, and M. A. McClure. Hidden markov models of biological
primary sequence information. PNAS, 91(3):1059?1063, 1994.
[10] T. Jaakkola, M. Diekhans, and D. Haussler. A discriminative framework for detecting remote
protein homologies. Journal of Computational Biology, 2000.
[11] T. Jaakkola, M. Diekhans, and D. Haussler. Using the fisher kernel method to detect remote
protein homologies. In Proceedings of the Seventh International Conference on Intelligent
Systems for Molecular Biology, pages 149?158. AAAI Press, 1999.
[12] V. N. Vapnik. Statistical Learning Theory. Springer, 1998.
[13] N. Cristianini and J. Shawe-Taylor. An Introduction to Support Vector Machines. Cambridge,
2000.
[14] C. Leslie, E. Eskin, and W. S. Noble. The spectrum kernel: A string kernel for SVM protein
classification. Proceedings of the Pacific Biocomputing Symposium, 2002.
[15] A. G. Murzin, S. E. Brenner, T. Hubbard, and C. Chothia. SCOP: A structural classification
of proteins database for the investigation of sequences and structures. Journal of Molecular
Biology, 247:536?540, 1995.
[16] M. Sagot. Spelling approximate or repeated motifs using a suffix tree. Lecture Notes in Computer Science, 1380:111?127, 1998.
[17] G. Pavesi, G. Mauri, and G. Pesole. An algorithm for finding signals of unknown length in DNA
sequences. Bioinformatics, 17:S207?S214, July 2001. Proceedings of the Ninth International
Conference on Intelligent Systems for Molecular Biology.
[18] S. Henikoff and J. G. Henikoff. Embedding strategies for effective use of information from
multiple sequence alignments. Protein Science, 6(3):698?705, 1997.
[19] S. L. Salzberg. On comparing classifiers: Pitfalls to avoid and a recommended approach. Data
Mining and Knowledge Discovery, 1:371?328, 1997.
[20] C. Watkins. Dynamic alignment kernels. Technical report, UL Royal Holloway, 1999.
[21] D. Haussler. Convolution kernels on discrete structure. Technical report, UC Santa Cruz, 1999.
[22] Huma Lodhi, John Shawe-Taylor, Nello Cristianini, and Chris Watkins. Text classification using
string kernels. Preprint.
| 2179 |@word version:1 norm:2 lodhi:1 mers:7 willing:1 additively:1 gish:1 substitution:1 series:1 score:23 selecting:2 prefix:5 existing:1 current:5 comparing:1 john:1 cruz:1 designed:2 plot:13 update:2 generative:5 leaf:5 pointer:3 eskin:2 detecting:1 node:12 traverse:1 simpler:1 zhang:2 phylogenetic:1 along:1 ucsc:1 symposium:1 descendant:1 combine:1 baldi:1 introduce:1 blast:5 pairwise:2 mpg:1 isi:1 multi:1 decreasing:1 pitfall:1 little:1 becomes:1 provided:1 string:15 differing:1 finding:1 exactly:1 classifier:11 scaled:1 grant:2 planck:1 positive:17 eleazar:1 local:1 oxford:1 path:5 approximately:1 signed:1 might:1 hmms:2 range:1 trie:1 mer:17 practical:2 acknowledgment:1 practice:2 recursive:1 area:1 remotely:1 significantly:3 pre:1 protein:27 selection:2 www:3 fruitful:1 map:10 lexical:1 baum:1 murzin:1 welch:1 assigns:1 haussler:5 dbi:1 pull:1 retrieve:1 embedding:1 coordinate:7 variation:1 target:2 suppose:1 massive:1 us:1 expensive:2 satisfying:1 labeled:2 database:7 preprint:1 capture:2 worst:2 calculate:2 thousand:1 stafford:1 remote:13 grundy:1 complexity:1 cristianini:2 traversal:3 gapped:1 dynamic:1 depend:1 serve:1 upon:1 efficiency:2 basis:1 represented:2 chapter:1 alphabet:3 train:5 fast:2 describe:1 effective:2 neighborhood:2 outside:1 whose:1 encoded:1 quite:2 widely:1 otherwise:2 sequence:42 myers:1 henikoff:2 product:2 maximal:1 combining:1 parent:1 empty:1 depending:1 andrew:1 measured:1 krogh:1 dividing:1 implemented:3 c:4 differ:1 generalization:1 investigation:1 biological:5 probable:1 mapping:1 pointing:1 achieves:1 fh:6 label:1 hubbard:1 tool:1 mclachlan:1 avoid:1 jaakkola:8 ax:1 focus:1 improvement:1 rank:1 likelihood:3 detect:2 motif:2 suffix:2 bairoch:1 entire:2 hidden:7 interested:1 germany:1 overall:2 classification:13 html:1 issue:1 sjolander:1 plan:1 biologist:1 uc:1 equal:2 saving:1 having:1 washington:2 lipman:2 biology:10 represents:2 look:2 noble:5 distantly:1 report:4 intelligent:3 beck:1 william:2 maintain:1 friedman:1 detection:15 mining:1 alignment:8 held:1 chain:2 tree:14 indexed:2 taylor:2 joyce:1 instance:11 soft:1 wb:3 modeling:1 salzberg:1 leslie:2 cost:1 subset:1 entry:3 successful:3 seventh:1 considerably:1 prosite:1 fundamental:1 international:3 michael:1 aaai:2 account:1 de:1 scop:9 wsn:1 includes:1 explicitly:1 unannotated:2 root:3 jason:1 lot:1 contribution:2 publicly:1 acid:5 efficiently:3 miller:2 substring:3 reach:1 inexpensive:2 associated:1 psi:4 timate:1 dataset:3 knowledge:1 ut:1 improves:1 eddy:1 sophisticated:1 actually:1 ta:6 supervised:1 follow:1 methodology:1 though:1 just:1 mian:1 defines:1 indicated:1 believe:1 grows:1 name:1 homologs:4 contain:1 homology:17 normalized:2 true:1 brown:1 requiring:1 during:1 bonferroni:1 rooted:1 compbio:2 criterion:2 m5:1 eggert:1 performs:3 jrl:1 recently:1 nih:1 functional:1 extend:2 m1:1 significant:1 cambridge:1 similarly:1 shawe:2 fingerprint:1 similarity:5 add:1 belongs:1 altschul:2 store:2 certain:1 outperforming:1 kfh:1 binary:1 seen:1 additional:1 redundant:1 recommended:1 signal:1 semi:1 branch:3 multiple:4 july:1 pnas:2 exceeds:1 technical:2 match:1 calculation:2 mcclure:1 christina:1 molecular:6 visit:1 prediction:2 involving:1 basic:1 kernel:70 represent:2 normalization:1 sometimes:1 want:2 source:2 unlike:1 member:2 structural:2 enough:1 chothia:1 gave:1 identified:1 inner:2 idea:2 diekhans:2 motivated:1 ul:1 useful:3 generally:1 santa:1 involve:1 svms:3 dna:2 http:1 nsf:1 dotted:2 disjoint:1 per:1 discrete:1 zv:1 four:1 threshold:2 achieving:1 backward:1 graph:2 sum:1 year:1 enforced:1 family:20 almost:1 distinguish:1 fold:2 quadratic:2 g:1 strength:1 occur:1 precisely:1 performing:2 relatively:1 bdc:12 department:3 pacific:1 combination:2 mismatching:3 slightly:1 character:2 sam:5 biologically:1 computationally:1 previously:1 discus:1 count:2 available:2 experimentation:1 occurrence:2 gribskov:1 encounter:1 original:1 eeskin:1 include:3 running:1 remaining:1 calculating:2 giving:1 build:1 added:1 print:1 occurs:1 strategy:1 primary:1 spelling:1 diagonal:1 evolutionary:2 gradient:1 thank:1 simulated:1 hmm:3 evenly:1 chris:2 nello:1 tuebingen:2 consensus:1 chauvin:1 length:17 difficult:1 potentially:3 holding:1 negative:3 implementation:1 unknown:1 perform:4 nucleic:3 convolution:1 markov:9 datasets:5 benchmark:2 finite:1 waterman:1 defining:1 ninth:1 community:2 schaffer:1 david:2 pair:1 connection:1 learned:1 huma:1 assembled:1 below:2 pattern:2 mismatch:55 usually:1 flower:1 program:1 max:1 wz:7 memory:1 including:2 royal:1 power:1 natural:1 rely:1 representing:1 scheme:1 improve:1 axis:4 madden:1 columbia:6 ues:1 nir:1 formerly:1 prior:1 text:2 discovery:1 val:1 relative:2 eisenberg:1 lecture:1 interesting:1 generation:1 degree:1 storing:2 supported:2 free:1 institute:1 fall:1 fifth:1 superfamily:3 fg:2 sparse:1 curve:1 depth:4 calculated:2 valid:5 genome:1 rich:1 kz:6 forward:1 unweighted:1 qualitatively:1 counted:1 ec:1 approximate:1 preferred:1 status:1 roc50:4 investigating:1 summing:1 discriminative:7 spectrum:14 subsequence:4 search:4 iterative:2 tailed:1 expanding:2 hunkapiller:1 contributes:2 cl:1 domain:3 profile:5 repeated:1 amino:2 roc:19 explicit:2 concatenating:1 watkins:3 weighting:1 third:1 down:2 showing:2 sift:1 symbol:2 list:1 explored:1 svm:44 false:3 vapnik:1 illustrates:1 occurring:3 margin:2 simply:1 adjustment:1 partially:1 springer:1 corresponds:1 weston:2 shared:1 fisher:29 considerable:1 change:1 soe:1 included:1 brenner:1 called:4 total:3 pas:1 e:1 experimental:1 indicating:1 holloway:1 internal:3 support:6 biocomputing:1 bioinformatics:1 incorporate:1 tested:2 |
1,295 | 218 | 750
Koch, Bair, Harris, Horiuchi, Hsu and Luo
Real- Time Computer Vision and Robotics
Using Analog VLSI Circuits
Christof Koch
Wyeth Bair
John G. Harris
Timothy Horiuchi
Andrew Hsu
Jin Luo
Computation and Neural Systems Program
Caltech 216-76
Pasadena, CA 91125
ABSTRACT
The long-term goal of our laboratory is the development of analog
resistive network-based VLSI implementations of early and intermediate vision algorithms. We demonstrate an experimental circuit for smoothing and segmenting noisy and sparse depth data
using the resistive fuse and a 1-D edge-detection circuit for computing zero-crossings using two resistive grids with different spaceconstants. To demonstrate the robustness of our algorithms and
of the fabricated analog CMOS VLSI chips, we are mounting these
circuits onto small mobile vehicles operating in a real-time, laboratory environment.
1
INTRODUCTION
A large number of computer vision algorithms for finding intensity edges, computing motion, depth, and color, and recovering the 3-D shapes of objects have been
developed within the framework of minimizing an associated "energy" functional.
Such a variational formalism is attractive because it allows a priori constraints
to be explicitly stated. The single most important constraint is that the physical
processes underlying image formation, such as depth, orientation and surface reflectance, change slowly in space. For instance, the depths of neighboring points on
a surface are usually very similar. Standard regularization algorithms embody this
smoothness constraint and lead to quadratic variational functionals with a unique
global minimum (Poggio, Torre, and Koch, 1985). These quadratic functionals
Real-Time Computer Vision and Robotics Using Analog VLSI Circuits
G
G
G
G
Rl
G
G
Rl
(a)
3.1V
Node
Voltage
(b)
3.0V
I? I?
?
1
1
Edge
Output
?
2
345
Photoreceptor
6
o
7
Figure 1: (a) shows the schematic of the zero-crossing chip. The phototransistors
logarithmically map light intensity to voltages that are applied via a conductance G
onto the nodes of two linear resistive networks. The network resistances Rl and R2
can be arbitrarily adjusted to achieve different space-constants. Transconductance
amplifiers compute the difference of the smoothed network node voltages and report
a current proportional to that difference. The sign of current then drives exclusive-or
circuitry (not shown) between each pair of neighboring pixels. The final output is
a binary signal indicating the positions of the zero-crossings. The linear network
resistances have been implemented using Mead's saturating resistor circuit (Mead,
1989), and the vertical resistors are implemented with transconductance followers.
(b) shows the measured response of a seven-pixel version of the chip to a bright
background with a shadow cast across the middle three photoreceptors. The circles
and triangles show the node voltages on the resistive networks with the smaller and
larger space-constants, respectively. Edges are indicated by the binary output (bar
chart at bottom) corresponding to the locations of zero-crossings.
751
752
Koch, Bair, Harris, Horiuchi, Hsu and Luo
can be mapped onto linear resistive networks, such that the stationary voltage distribution, corresponding to the state of least power dissipation, is equivalent to
the solution of the variational functional (Horn, 1974; Poggio and Koch , 1985).
Smoothness breaks down, however, at discontinuities caused by occlusions or differences in the physical processes underlying image formation (e.g., different surface
reflectance properties). Detecting these discontinuities becomes crucial, not only
because otherwise smoothness is incorrectly applied but also because the locations
of discontinuities are often required for further image analysis and understanding.
We describe two different approaches for finding discontinuities in early vision: (1)
a 1-D edge-detection circuit for computing zero-crossings using two resistive grids
with different space-constants, and (2) a 20 by 20 pixel circuit for smoothing and
segmenting noisy and sparse depth data using the resistive fuse.
Finally, while successfully demonstrating a highly integrated circuit on a stationary
laboratory bench under controlled conditions is already a tremendous success, this
is not the environment in which we ultimately intend them to be used. The jump
from a sterile, well-controlled, and predictable environment such as that of the
laboratory bench to a noisy and physically demanding environment of a mobile
robot can often spell out the true limits of a circuit's robustness. In order to
demonstrate the robustness and real-time performance of these circuits, we have
mounted two such chips onto small toy vehicles.
2
AN EDGE DETECTION CIRCUIT
The zero-crossings of the Laplacian of the Gaussian, V 2 G, are often used for detecting edges. Marr and Hildreth (1980) discovered that the Mexican-hat shape
of the V2G operator can be approximated by the difference of two Gaussians
(DOG). In this spirit, we have built a chip that takes the difference of two resistivenetwork smoothings of photoreceptor input and finds the resulting zero-crossings.
The Green's function of the resistive network, a decaying exponential, differs from
the Gaussian, but simulations with digitized camera images have shown that the
difference of exponentials (DOE) gives results nearly as good as the DOG. Furthermore, resistive nets have a natural implementation in silicon, while implementing
the Gaussian is cumbersome.
The circuit, Figure la, uses two independent resistive networks to smooth the voltages supplied by logarithmic photoreceptors. The voltages on the two networks are
subtracted and exclusive-or circuitry (not shown) is used to detect zero-crossings. In
order to facilitate thresholding of edges, an additional current is computed at each
node indicating the strength of the zero-crossing. This is particularly important
for robust real-world performance where there will be many small (in magnitude
of slope) zero-crossings due to noise. Figure 1b shows the measured response of a
seven-pixel version of the chip to a bright background with a shadow cast across the
middle three photoreceptors. Subtracting the two network voltage traces shown at
the top, we find two zero-crossings, which the chip correctly identifies in the binary
output shown at the bottom.
Real-Time Computer Vision and Robotics Using Analog VLSI Circuits
~OJ
V
--
00
V- ~ ~,
_
J
./
(a)
........
vr-.....
,-
--
~
c::J-
I~
/,:1
I
...
G-
\' -
-
~f;j
-It\.
I:::j-
~I..
1 ~
2u 2 ~
dij
c::J-
1.
~
~
o0
/
~"""I
0 I Il
HRES
300
I
(nA)
(b)
O+-__~____-~V~T~~~________~____
-30~0.5
~V
0.0
(Volts)
0.5
Figure 2: (a) Schematic diagram for the 20 by 20 pixel surface interpolation and
smoothing chip. A rectangular mesh of resistive fuse elements (shown as rectangles)
provide the smoothing and segmentation ability of the network. The data are given
as battery values dij with the conductance G connecting the battery to the grid set
to G = 1/2u 2 , where u 2 is the variance of the additive Gaussian noise assumed to
corrupt the data. (b) Measured current-voltage relationship for different settings
of the resistive fuse. For a voltage of less than VT across this two-terminal device,
the circuit acts as a resistor with conductance A. Above VT, the current is either
abruptly set to zero (binary fuse) or smoothly goes to zero (analog fuse). We can
continuously vary the I-V curve from the hyperbolic tangent of Mead's saturating
resistor (HRES) to that of an analog fuse (Fig. 2b), effectively implementing a
continuation method for minimizing the non-convex functional. The I-V curve of a
binary fuse is also illustrated.
753
754
Koch, Bair, Harris, Horiuchi, Hsu and Luo
3
A CIRCUIT FOR SMOOTHING AND SEGMENTING
Many researchers have extended regularization theory to include discontinuities.
Let us consider the problem of interpolating noisy and sparse 1-D data (the 2-D
generalization is straightforward), where the depth data di is given on a discrete
grid. Associated with each lattice point is the value of the recovered surface Ii
and a binary line discontinuity Ii. When the surface is expected to be smooth
(with a first-order, membrane-type stabilizer) except at isolated discontinuities, the
functional to be minimized is given by:
J(f, I) = A~(fi+l - 1i)2(1 -Ii)
+ 2!2 ~(di -
I
I
1i)2 + a
~ Ii
(1)
I
where (]'2 is the variance of the additive Gaussian noise process assumed to corrupt
the data di, and A and a are free parameters. The first term implements the
piecewise smooth constraint: if all variables, with the exception of Ii, Ii+l, and Ii,
are held fixed and A(fi+l - h)2 < a, it is "cheaper" to pay the price A(fi+l - h)2
and set Ii = 0 than to pay the larger price a; if the gradient becomes too steep,
Ii = 1, and the surface is segmented at that location. The second term, with the
sum only including those locations i where data exist, forces the surface I to be
close to the measured data d. How close depends on the estimated magnitude of
the noise, in this case on (]'2. The final surface I is the one that best satisfies the
conflicting demands of piecewise smoothness and fidelity on the measured data.
To minimize the 2-D generalization of eq. (1), we map the functional J onto the
circuit shown in Fig. 2a such that the stationary voltage at every gridpoint then
corresponds to hi. The cost functional J is interpreted as electrical co-content,
the generalization of power for nonlinear networks. We designed a two-terminal
nonlinear device, which we call a resistive fuse, to implement piecewise smoothness
(Fig. 2b). If the magnitude of the voltage drop across the device is less than
VT = (a/A)1/2, the fuse acts as a linear resistor with conductance A. If VT is
exceeded, however, the fuse breaks and the current goes to zero. The operation of
the fuse is fully reversible. We built a 20 by 20 pixel fuse network chip and show
its segmentation and smoothing performance in Figure 3.
4
AUTONOMOUS VEHICLES
Our goal-beyond the design and fabrication of analog resistive-network chips-is
to build mobile testbeds for the evaluation of chips as well as to provide a systems
perspective on the usefulness of certain vision algorithms. Due to the small size
and power requirements of these chips, it is possible to utilize the vast resource of
commercially available toy vehicles. The advantages of toy cars over robotic vehicles
built for research are their low cost, ease of modification, high power-to-weight ratio,
availability, and inherent robustness to the real-world. Accordingly, we integrated
two analog resistive-network chips designed and built in Mead's laboratory onto
small toy cars controlled by a digital microprocessor (see Figure 4).
Real-Time Computer Vision and Robotics Using Analog VLSI Circuits
(c)
(b)
<a)
M
M
M
'.1
'.1
?
1.1
1.0
'.1
'.0
~
12 10
"Noel.?Number
(d)
20
( e)
1.0
...1""'' \
.".'
1.1
1.1
0
1.1
12 10
"Noel.? Number
20
( f)
12 10
"Noel.?Number
20
Figure 3: Experimental data from the fuse network chip. We use as input data a
tower (corresponding to dij = 3.0 V) rising from a plane (corresponding to 2.0 V)
with superimposed Gaussian noise. (a) shows the input with the variance of the
noise set to 0.2 V, (b) the voltage output using the fuse configured as a saturating
resistance, and (c) the output when the fuse elements are activated. (d), (e),
and (f) illustrate the same behavior along a horizontal slice across the chip for
(12
0.4 V. We used a hardware deterministic algorithm of varying the fuse I-V
curve of the saturating resistor to that of the analog fuse (following the arroW in
Fig. 2b) as well as increasing the conductance A. This algorithm is closely related
to other deterministic approximations based on continuation methods or a Mean
Field Theory approach (Koch, Marroquin, and Yuille, 1986; Blake and Zisserman,
1987; Geiger and Girosi, 1989). Notice that the amplitude of the noise in the last
case (40% of the amplitude of the voltage step) is so large that a single filtering
step on the input (d) will fail to detect the tower. Cooperativity and hysteresis
are required for optimal performance. Notice the "bad" pixel in the middle of the
tower (in c). Its effect is localized, however, to a single element.
=
755
| 218 |@word middle:3 version:2 rising:1 simulation:1 current:6 recovered:1 luo:4 follower:1 john:1 mesh:1 additive:2 girosi:1 shape:2 designed:2 drop:1 mounting:1 stationary:3 device:3 accordingly:1 plane:1 detecting:2 node:5 location:4 along:1 resistive:16 expected:1 behavior:1 embody:1 terminal:2 increasing:1 becomes:2 underlying:2 circuit:18 interpreted:1 developed:1 finding:2 fabricated:1 every:1 act:2 christof:1 segmenting:3 limit:1 mead:4 interpolation:1 hres:2 co:1 ease:1 unique:1 horn:1 camera:1 implement:2 differs:1 hyperbolic:1 onto:6 close:2 operator:1 equivalent:1 map:2 deterministic:2 go:2 straightforward:1 convex:1 rectangular:1 marr:1 autonomous:1 us:1 crossing:11 logarithmically:1 approximated:1 particularly:1 element:3 bottom:2 electrical:1 environment:4 predictable:1 battery:2 ultimately:1 yuille:1 triangle:1 chip:15 horiuchi:4 describe:1 cooperativity:1 formation:2 larger:2 otherwise:1 ability:1 noisy:4 final:2 advantage:1 net:1 subtracting:1 neighboring:2 achieve:1 requirement:1 cmos:1 object:1 illustrate:1 andrew:1 measured:5 eq:1 recovering:1 implemented:2 shadow:2 closely:1 torre:1 implementing:2 generalization:3 adjusted:1 koch:7 blake:1 circuitry:2 vary:1 early:2 successfully:1 gaussian:6 mobile:3 voltage:14 varying:1 superimposed:1 detect:2 integrated:2 pasadena:1 vlsi:6 pixel:7 fidelity:1 orientation:1 priori:1 development:1 smoothing:6 field:1 testbeds:1 nearly:1 commercially:1 minimized:1 report:1 piecewise:3 inherent:1 cheaper:1 occlusion:1 amplifier:1 detection:3 conductance:5 highly:1 evaluation:1 light:1 activated:1 held:1 edge:8 poggio:2 circle:1 isolated:1 instance:1 formalism:1 lattice:1 cost:2 usefulness:1 fabrication:1 dij:3 too:1 connecting:1 continuously:1 na:1 slowly:1 toy:4 availability:1 hysteresis:1 configured:1 explicitly:1 caused:1 depends:1 vehicle:5 break:2 wyeth:1 decaying:1 slope:1 minimize:1 bright:2 chart:1 il:1 variance:3 gridpoint:1 drive:1 researcher:1 cumbersome:1 energy:1 associated:2 di:3 hsu:4 color:1 car:2 segmentation:2 amplitude:2 marroquin:1 exceeded:1 response:2 zisserman:1 furthermore:1 horizontal:1 nonlinear:2 reversible:1 hildreth:1 indicated:1 facilitate:1 effect:1 true:1 spell:1 regularization:2 volt:1 laboratory:5 illustrated:1 attractive:1 demonstrate:3 motion:1 dissipation:1 image:4 variational:3 fi:3 functional:6 physical:2 rl:3 analog:11 silicon:1 smoothness:5 grid:4 robot:1 operating:1 surface:9 perspective:1 certain:1 binary:6 arbitrarily:1 success:1 vt:4 caltech:1 minimum:1 additional:1 signal:1 ii:9 smooth:3 segmented:1 long:1 controlled:3 schematic:2 laplacian:1 vision:8 physically:1 robotics:4 background:2 diagram:1 crucial:1 spirit:1 call:1 intermediate:1 stabilizer:1 bair:4 o0:1 abruptly:1 resistance:3 hardware:1 continuation:2 supplied:1 exist:1 notice:2 sign:1 estimated:1 correctly:1 discrete:1 demonstrating:1 utilize:1 rectangle:1 vast:1 fuse:18 sum:1 geiger:1 hi:1 pay:2 quadratic:2 strength:1 constraint:4 transconductance:2 membrane:1 across:5 smaller:1 modification:1 resource:1 fail:1 available:1 gaussians:1 operation:1 subtracted:1 robustness:4 hat:1 top:1 include:1 reflectance:2 build:1 intend:1 already:1 exclusive:2 gradient:1 mapped:1 seven:2 tower:3 relationship:1 ratio:1 minimizing:2 steep:1 trace:1 stated:1 implementation:2 design:1 vertical:1 jin:1 incorrectly:1 extended:1 digitized:1 discovered:1 smoothed:1 intensity:2 pair:1 cast:2 required:2 dog:2 conflicting:1 tremendous:1 discontinuity:7 beyond:1 bar:1 usually:1 program:1 built:4 green:1 including:1 oj:1 power:4 demanding:1 natural:1 force:1 identifies:1 understanding:1 tangent:1 fully:1 proportional:1 mounted:1 filtering:1 localized:1 digital:1 thresholding:1 corrupt:2 last:1 free:1 sparse:3 slice:1 curve:3 depth:6 world:2 jump:1 functionals:2 global:1 robotic:1 photoreceptors:3 assumed:2 robust:1 ca:1 interpolating:1 microprocessor:1 arrow:1 noise:7 fig:4 vr:1 position:1 resistor:6 exponential:2 down:1 bad:1 r2:1 phototransistors:1 effectively:1 magnitude:3 demand:1 smoothly:1 logarithmic:1 timothy:1 saturating:4 corresponds:1 satisfies:1 harris:4 goal:2 noel:3 price:2 content:1 change:1 except:1 mexican:1 experimental:2 la:1 photoreceptor:2 indicating:2 exception:1 bench:2 |
1,296 | 2,180 | Automatic Alignment of Local Representations
Yee Whye Teh and Sam Roweis
Department of Computer Science, University of Toronto
ywteh,roweis @cs.toronto.edu
Abstract
We present an automatic alignment procedure which maps the disparate
internal representations learned by several local dimensionality reduction
experts into a single, coherent global coordinate system for the original
data space. Our algorithm can be applied to any set of experts, each
of which produces a low-dimensional local representation of a highdimensional input. Unlike recent efforts to coordinate such models by
modifying their objective functions [1, 2], our algorithm is invoked after
training and applies an efficient eigensolver to post-process the trained
models. The post-processing has no local optima and the size of the system it must solve scales with the number of local models rather than the
number of original data points, making it more efficient than model-free
algorithms such as Isomap [3] or LLE [4].
1 Introduction: Local vs. Global Dimensionality Reduction
Beyond density modelling, an important goal of unsupervised learning is to discover compact, informative representations of high-dimensional data. If the data lie on a smooth low
dimensional manifold, then an excellent encoding is the coordinates internal to that manifold. The process of determining such coordinates is dimensionality reduction. Linear
dimensionality reduction methods such as principal component analysis and factor analysis are easy to train but cannot capture the structure of curved manifolds. Mixtures of these
simple unsupervised models [5, 6, 7, 8] have been used to perform local dimensionality
reduction, and can provide good density models for curved manifolds, but unfortunately
such mixtures cannot do dimensionality reduction. They do not describe a single, coherent low-dimensional coordinate system for the data since there is no pressure for the local
coordinates of each component to agree.
Roweis et al [1] recently proposed a model which performs global coordination of local
coordinate systems in a mixture of factor analyzers (MFA). Their model is trained by maximizing the likelihood of the data, with an additional variational penalty term to encourage
the internal coordinates of the factor analyzers to agree. While their model can trade off
modelling the data and having consistent local coordinate systems, it requires a user given
trade-off parameter, training is quite inefficient (although [2] describes an improved training algorithm for a more constrained model), and it has quite serious local minima problems
(methods like LLE [4] or Isomap [3] have to be used for initialization).
In this paper we describe a novel, automatic way to align the hidden representations used by
each component of a mixture of dimensionality reducers into a single global representation
of the data throughout space. Given an already trained mixture, the alignment is achieved
by applying an eigensolver to a matrix constructed from the internal representations of the
mixture components. Our method is efficient, simple to implement, and has no local optima
in its optimization nor any learning rates or annealing schedules.
2 The Locally Linear Coordination Algorithm
!
#
%$ * '
$,+.'
/ $ * '
$ 10
Suppose we have a set of data points given by the rows of
from
a -dimensional space, which we assume are sampled from a
dimensional manifold. We approximate the manifold coordinates using images
in a
dimensional embedding space. Suppose also that we have already trained, or have been
given, a mixture of local dimensionality reducers. The th reducer produces a dimensional internal representation
for data point
as well as a ?responsibility?
describing how reliable the th reducer?s representation of
is. These satisfy
and can be obtained, for example, using a gating network in a mixture of experts, or the
posterior probabilities in a probabilistic network. Notice that the manifold coordinates and
internal representations need not have the same number of dimensions.
"
# &('
$
)'
Given the data, internal representations, and responsibilities, our algorithm automatically
aligns the various hidden representations into a single global coordinate system. Two key
ideas motivate the method. First, to use a convex cost function whose unique minimum is
attained at the desired global coordinates. Second, to restrict the global coordinates
to
depend on the data
only through the local representations
and responsibilities
,
thereby leveraging the structure of the mixture model to regularize and reduce the effective
size of the optimization problem. In effect, rather than working with individual data points,
we work with large groups of points belonging to particular submodels.
'
3 '$
* ' '2$
& '
$
' & '
$ * '
$ & '
$
?A@ B B
3
* '2$%9 $ & '
$;: 4<5$>= 87 $ 7BDC * '
$>E '
$ 4 $ B F7>GH ' G 4 G B
J J 9LK # = HM5 ' G N*O'
$P& '
$ 4 G N4 $
3 $ E '2B $ K
& '
$ E '
5 $ Q0
J 9LK # = J 9LK R# =
J B J 9K R# =
0
RS%A3 " :,/ G $ $ B
G
J
3
I
HT' F*O'
$ E '
$ I
4 .4 $ 3
in terms of
and
. Given an input
We first parameterize the global coordinates
, each local model infers its internal coordinates
and then applies a linear projection
and offset
to these to obtain its guess at the global coordinates. The final global
is obtained by averaging the guesses using the responsibilities as weights:
coordinates
' 465$
' 87 $
F8I 3
B
4$ K
9LK # = 9LK R# =
(1)
(2)
where
is the th column of
,
is the th entry of
, and
is a bias.
This process is described in figure 1. To simplify our calculations, we have vectorized the
indices
into a single new index
, where
is an invertible mapping from
to
. For compactness, we will write
.
the domain of
Now define the matrices and as
and the th row of as
. Then
(1) becomes a system of linear equations (2) with fixed and unknown parameters .
responsibilities
r nk
unj
high?dimensional
data
xn
alignment
parameters
lj
yn
global
coordinates
z nk
local dimensionality
reduction models
Responsibility?weighted
local representations
local coordinates
Figure 1: Obtaining global coordinates from data via responsibility-weighted local coordinates.
''
B
The key assumption, which we have emphasized by re-expressing
above, is that the
mapping between the local representations and the global coordinates
is linear in each
of
,
and the unknown parameters . Crucially, however, the mapping between the
original data
and the images
is highly non-linear since it depends on the multiplication of responsibilities and internal coordinates which are in turn non-linearly related to
the data
through the inference procedure of the mixture model.
& '
$ * '
$ '
'
9 =
3
'
4$
3 9 9 3 =
9 = =
3
9 = -
I I
3
3 ? ?
3 I 3
I ?
We now consider determining according to some given cost function
. For this we
advocate using a convex
. Notice that since is linear in ,
is convex in
as well, and there is a unique optimum that can be computed efficiently using a variety of
methods. This is still true if we also have feasible convex constraints
on . The
case where the cost and constraints are both quadratic is particularly appealing since we
can use an eigensolver to find the optimal . In particular suppose and are matrices
defining the cost and constraints, and let
and
. This gives:
9 = 3 3
3 I 3 I
3
8 FI I
9 = -
(3)
where
is the trace operator. The matrices
and
are typically obtained from the
original data and summarize the essential geometries among them. The solution to the
constrained minimization above is given by the smallest generalized eigenvectors with
. In particular the columns of are given by these generalized eigenvectors.
3
Below, we investigate a cost function based on the Locally Linear Embedding (LLE) algorithm of Roweis and Saul [4]. We call the resulting algorithm Locally Linear Coordination (LLC). The idea of LLE is to preserve the same locally linear relationships between
the original data points
and their counterparts . We identify for each point
its
nearest-neighbours
and then minimize
'
'
' '
9 , = F7 ' '
/ !#"#$ ' % ( 9 &
' = 9 (
' =
/ ! " $ ' 0
$ '
)
'
9 = * ( 9 &
' = 9 (
' =
+ 0 7 ' ' +-0 0, + 0 7 ' ' ' + 0 *
?
0,
0 3
9 . 3 = % ( I 9 ( 3
= 9 (3
' 3 = I 3 3 M* /O 3 3
0 , MI I I 0 % ?
8FI 9 (
' = 9
= I + 0 I I
(4)
subject to the constraints
. The weights are unique1
with respect to
and can be solved for efficiently using constrained least squares (since solving for
is
decoupled across ). The weights summarize the local geometries relating the data points
to their neighbours, hence to preserve these relationships among the coordinates
we
arrange to minimize the same cost
(5)
but with respect to instead. is invariant to translations and rotations of , and scales as
we scale . In order to break these degeneracies we enforce the following constraints:
(6)
where is a vector of ?s. For this choice, the cost function and constraints above become:
(7)
(8)
with cost and constraint matrices
1
(9)
1
In the unusual case where the number of neighbours is larger than the dimensionality of the data
, simple regularization of the norm of the weights once again makes them unique.
0 , I 3
5 I /,0
5
8 0(
As shown previously, the solution to this problem is given by the smallest generalized
. To satisfy
, we need to find eigenvectors
eigenvectors with
that are orthogonal to the vector
. Fortunately,
is the smallest generalized
eigenvector, correspondingto
an eigenvalue of 0. Hence the solution to the problem is
given by the
to
smallest generalized eigenvectors instead.
S ' ? 9 : 0 =
LLC Alignment Algorithm:
Using data
, compute local linear reconstruction weights
$ '
using (4).
& '
$ '
Train or receive a pre-trained mixture of local dimensionality reducers.
Apply this mixture to , obtaining a local representation
and
responsibility
for each submodel and each data point .
* '
$
Form the matrix
I
with
GH ' * '
$>E '
B #$
: 0
BS : 0
4 F$ I 3
and calculate
0(
Find the eigenvectors corresponding to the smallest
of the generalized eigenvalue system
.
3
J
3
nd
Let be a matrix with columns formed by the
Return the th row of as alignment weight .
Return the global manifold coordinates as
" : / $$
to
and
from (9).
eigenvalues
st
eigenvectors.
.
Note that the edge size of the matrices and whose generalized eigenvectors we seek
which scales with the number of components and dimensions of the local
is
representations but not with the number of data points . As a result, solving for the
alignment weights is much more efficient than the original LLE computation (or those
of Isomap) which requires solving an eigenvalue system of edge size . In effect, we
have leveraged the mixture of local models to collapse large groups of points together and
worked only with those groups rather than the original data points. Notice however that
the computation of the weights
still requires determining the neighbours of the original
in the worse case.
data points, which scales as
+
9 + =
#
+
*P4 5$$
3$
Coordination with LLC also yields a mixture of noiseless factor analyzers over the global
coordinate space , with the th factor analyzer having mean
and factor loading
.
Given any global coordinates , we can infer the responsibilities and the posterior means
over the latent space of each factor analyzer. If our original local dimensionality reducers also supports computing from and , we can now infer the high dimensional mean
data point which corresponds to the global coordinates . This allows us to perform operations like visualization and interpolation using the global coordinate system. This is the
method we used to infer the images in figures 4 and 5 in the next section.
&$
*$ &$
3 Experimental Results using Mixtures of Factor Analyzers
The alignment computation we have described is applicable to any mixture of local dimensionality reducers. In our experiments, we have used the most basic such model: a mixture
of factor analyzers (MFA) [8]. The th factor analyzer in the mixture describes a probabilistic linear mapping from a latent variable
to the data with additive Gaussian noise.
The model assumes that the data manifold is locally linear and it is this local structure that
is captured by each factor analyzer. The non-linearity in the data manifold is handled by
patching multiple factor analyzers together, each handling a locally linear region.
#
&$
MFAs are trained in an unsupervised way by maximizing the marginal log likelihood of
the observed data, and parameter estimation is typically done using the EM algorithm 2.
2
In our experiments, we initialized the parameters by drawing the means from the global covariance of the data and setting the factors to small random values. We also simplified the factor analyzers
to share the same spherical noise covariance
although this is not essential to the process.
A
B
C
D
Figure 2: LLC on the S curve (A). There are 14 factor analyzers in the mixture (B), each with 2 latent
dimensions. Each disk represents one of them with the two black lines being the factor loadings. After
alignment by LLC (C), the curve is successfully unrolled; it is also possible to retroactively align the
original data space models (D).
A
Figure 3: Unknotting the trefoil
B
curve. We generated 6000 noisy
points from the curve. Then we fit
an MFA with 30 components with
1 latent dimension each (A), but
aligned them in a 2D space (B).
We used 10 neighbours to reconstruct each data point.
&$
Since there is no constraint relating the various hidden variables , a MFA trained only
to maximize likelihood cannot learn a global coordinate system for the manifold that is
consistent across every factor analyzer. Hence this is a perfect model on which to apply
automatic alignment. Naturally, we use the mean of conditioned on the data (assuming
the th factor analyzer generated ) as the th local representation of , while we use the
posterior probability that the th factor analyzer generated as the responsibility.
#
#
#
&$
We illustrate LLC on two synthetic toy problems to give some intuition about how it works.
The first problem is the S curve given in figure 2(A). An MFA trained on 1200 points
sampled uniformly from the manifold with added noise (B) is able to model the linear
structure of the curve locally, however the internal coordinates of the factor analyzers are
not aligned properly. We applied LLC to the local representations and aligned them in a 2D
space (C). When solving for local weights, we used 12 neighbours to reconstruct each data
point. We see that LLC has successfully unrolled the S curve onto the 2D space. Further,
given the coordinate transforms produced by LLC, we can retroactively align the latent
spaces of the MFAs (D). This is done by determining directions in the various latent spaces
which get transformed to the same direction in the global space.
To emphasize the topological advantages of aligning representations into a space of higher
dimensionality than the local coordinates used by each submodel, we also trained a MFA
on data sampled from a trefoil curve, as shown in figure 3(A). The trefoil is a circle with a
knot in 3D. As figure 3(B) shows, LLC connects these models into a ring of local topology
faithful to the original data.
We applied LLC to MFAs trained on sets of real images believed to come from a complex
manifold with few degrees of freedom. We studied face images of a single person under
varying pose and expression changes and handwritten digits from the MNIST database.
After training the MFAs, we applied LLC to align the models. The face models were
aligned into a 2D space as shown in figure 4. The first dimension appears to describe
Figure 4: A map of reconstructions of the faces when the global coordinates are specified. Contours
describe the likelihood of the coordinates. Note that some reconstructions around the edge of the map
are not good because the model is extrapolating from the training images to regions of low likelihood.
A MFA with 20 components and 8 latent dimensions each is used. It is trained on 1965 images. The
weights
are calculated using 36 neighbours.
changes in pose, while the second describes changes in expression. The digit models were
aligned into a 3D space. Figure 5 (top) shows maps of reconstructions of the digits. The
first dimension appears to describe the slant of each digit, the second the fatness of each
digit, and the third the relative sizes of the upper to lower loops. Figure 5 (bottom) shows
how LLC can smoothly interpolate between any two digits. In particular, the first row
interpolates between left and right slanting digits, the second between fat and thin digits,
the third between thick and thin line strokes, and the fourth between having a larger bottom
loop and larger top loop.
4 Discussion and Conclusions
Previous work on nonlinear dimensionality reduction has usually emphasized either a parametric approach, which explicitly constructs a (sometimes probabilistic) mapping between
the high-dimensional and low-dimensional spaces, or a nonparametric approach which
merely finds low-dimensional images corresponding to high-dimensional data points but
without probabilistic models or hidden variables. Compared to the global coordination
model [1], the closest parametric approach to ours, our algorithm can be understood as post
coordination, in which the latent spaces are coordinated after they have been fit to data. By
decoupling the data fitting and coordination problems we gain efficiency and avoid local
optima in the coordination phase. Further, since we are just maximizing likelihood when
fitting the original mixture to data, we can use a whole range of known techniques to escape
local minima, and improve efficiency in the first phase as well.
On the nonparametric side, our approach can be compared to two recent algorithms, LLE
Figure 5: Top: maps of reconstructions of digits when two global coordinates are specified, and the
third integrated out. Left: st and nd coordinates specified; right: nd and rd . Bottom: Interpolating
between two digits using LLC. In each row, we interpolate between the upper leftmost and rightmost
digits. The LLC interpolants are spread out evenly along a line connecting the global coordinates of
the two digits. For comparison, we show the 20 training images whose coordinates are closest to the
line segment connecting those of the two digits at each side. A MFA with 50 components, each with
6 latent dimensions is used. It is trained on 6000 randomly chosen digits from the combined training
and test sets of 8?s in MNIST. The weights
were calculated using 36 neighbours.
[4] and Isomap [3]. The cost functions of LLE and Isomap are convex, so they do not
suffer from the local minima problems of earlier methods [9, 10], but these methods must
solve eigenvalue systems of size equal to the number of data points. (Although in LLE the
systems are highly sparse.) Another problem is neither LLE nor Isomap yield a probabilistic model or even a mapping between the data and embedding spaces. Compared to these
models (which are run on individual data points) LLC uses as its primitives descriptions
of the data provided by the individual local models. This makes the eigenvalue system to
be solved much smaller and as a result the computational cost of the coordination phase of
LLC is much less than that for LLE or Isomap. (Note that the construction of the eigenvalue
system still requires finding nearest neighbours for each point, which is costly.) Furthermore, if each local model describes a complete (probabilistic) mapping from data space
to its latent space, the final coordinated model will also describe a (probabilistic) mapping
from the whole data space to the coordinated embedding space.
Our alignment algorithm improves upon local embedding or density models by elevating
their status to full global dimensionality reduction algorithms without requiring any modifications to their training procedures or cost functions. For example, using mixtures of factor
analyzers (MFAs) as a test case, we show how our alignment method can allow a model
previously suited only for density estimation to do complex operations on high dimensional
data such as visualization and interpolation.
Brand [11] has recently proposed an approach, similar to ours, that coordinates local parametric models to obtain a globally valid nonlinear embedding function. Like LLC, his
?charting? method defines a quadratic cost function and finds the optimal coordination directly and efficiently. However, charting is based on a cost function much closer in spirit to
the original global coordination model and it instantiates one local model centred on each
training point, so its scaling is the same as that of LLE and Isomap. In principle, Brand?s
method can be extended to work with fewer local models and our alignment procedure can
be applied using the charting cost rather than the LLE cost we have pursued here.
Automatic alignment procedures emphasizes a powerful but often overlooked interpretation of local mixture models. Rather than considering the output of such systems to be a
single quantity, such as a density estimate or a expert-weighted regression, it is possible
to view them as networks which convert high-dimensional inputs into a vector of internal
coordinates from each submodel, accompanied by responsibilities. As we have shown, this
view can lead to efficient and powerful algorithms which allow separate local models to
learn consistent global representations.
Acknowledgments
We thank Geoffrey Hinton for inspiration and interesting discussions, Brendan Frey and
Yann LeCun for sharing their data sets, and the reviewers for helpful comments.
References
[1] S. Roweis, L. Saul, and G. E. Hinton. Global coordination of local linear models. In Advances
in Neural Information Processing Systems, volume 14, 2002.
[2] J. J. Verbeek, N. Vlassis, and B. Kr?ose. Coordinating principal component analysers. In Proceedings of the International Conference on Artificial Neural Networks, 2002.
[3] J. B. Tenenbaum, V. de Silva, and J. C. Langford. A global geometric framework for nonlinear
dimensionality reduction. Science, 290(5500):2319?2323, December 2000.
[4] S. Roweis and L. Saul. Nonlinear dimensionality reduction by locally linear embedding. Science, 290(5500):2323?2326, December 2000.
[5] K. Fukunaga and D. R. Olsen. An algorithm for finding intrinsic dimensionality of data. IEEE
Transactions on Computers, 20(2):176?193, 1971.
[6] N. Kambhatla and T. K. Leen. Dimension reduction by local principal component analysis.
Neural Computation, 9:1493?1516, 1997.
[7] M. E. Tipping and C. M. Bishop. Mixtures of probabilistic principal component analysers.
Neural Computation, 11(2):443?482, 1999.
[8] Z. Ghahramani and G. E. Hinton. The EM algorithm for mixtures of factor analyzers. Technical
Report CRG-TR-96-1, University of Toronto, Department of Computer Science, 1996.
[9] T. Kohonen. Self-organization and Associative Memory. Springer-Verlag, Berlin, 1988.
[10] C. Bishop, M. Svensen, and C. Williams. GTM: The generative topographic mapping. Neural
Computation, 10:215?234, 1998.
[11] M. Brand. Charting a manifold. This volume, 2003.
| 2180 |@word norm:1 loading:2 nd:3 disk:1 r:1 crucially:1 seek:1 covariance:2 pressure:1 thereby:1 tr:1 reduction:12 ours:2 rightmost:1 must:2 additive:1 informative:1 extrapolating:1 v:1 pursued:1 fewer:1 guess:2 generative:1 retroactively:2 toronto:3 along:1 constructed:1 become:1 advocate:1 fitting:2 nor:2 globally:1 spherical:1 automatically:1 considering:1 becomes:1 provided:1 discover:1 linearity:1 eigenvector:1 finding:2 every:1 fat:1 yn:1 understood:1 local:47 frey:1 encoding:1 interpolation:2 black:1 initialization:1 studied:1 collapse:1 range:1 unique:3 faithful:1 acknowledgment:1 lecun:1 implement:1 digit:14 procedure:5 projection:1 pre:1 get:1 onto:1 cannot:3 operator:1 applying:1 yee:1 map:5 reviewer:1 maximizing:3 primitive:1 williams:1 convex:5 submodel:3 regularize:1 his:1 embedding:7 coordinate:42 construction:1 suppose:3 user:1 us:1 particularly:1 database:1 observed:1 bottom:3 solved:2 capture:1 parameterize:1 calculate:1 region:2 trade:2 reducer:7 intuition:1 trained:12 motivate:1 depend:1 solving:4 segment:1 upon:1 efficiency:2 various:3 gtm:1 train:2 describe:6 effective:1 artificial:1 analyser:2 quite:2 whose:3 larger:3 solve:2 drawing:1 reconstruct:2 topographic:1 noisy:1 final:2 associative:1 advantage:1 eigenvalue:7 reconstruction:5 p4:1 kohonen:1 aligned:5 loop:3 roweis:6 description:1 optimum:4 produce:2 perfect:1 ring:1 illustrate:1 svensen:1 pose:2 nearest:2 c:1 come:1 direction:2 thick:1 modifying:1 slanting:1 crg:1 around:1 elevating:1 mapping:9 kambhatla:1 arrange:1 smallest:5 estimation:2 f7:2 applicable:1 coordination:12 successfully:2 weighted:3 minimization:1 gaussian:1 rather:5 avoid:1 varying:1 properly:1 modelling:2 likelihood:6 brendan:1 helpful:1 inference:1 lj:1 typically:2 integrated:1 compactness:1 hidden:4 transformed:1 among:2 constrained:3 marginal:1 equal:1 once:1 construct:1 having:3 represents:1 unsupervised:3 thin:2 report:1 simplify:1 serious:1 few:1 escape:1 randomly:1 neighbour:9 preserve:2 interpolate:2 individual:3 geometry:2 connects:1 phase:3 freedom:1 organization:1 highly:2 investigate:1 alignment:14 mixture:24 edge:3 encourage:1 closer:1 decoupled:1 orthogonal:1 initialized:1 desired:1 re:1 circle:1 column:3 earlier:1 cost:15 entry:1 synthetic:1 combined:1 st:2 density:5 person:1 international:1 probabilistic:8 off:2 invertible:1 together:2 connecting:2 again:1 leveraged:1 worse:1 expert:4 inefficient:1 return:2 toy:1 de:1 centred:1 accompanied:1 satisfy:2 coordinated:3 explicitly:1 depends:1 break:1 view:2 responsibility:12 minimize:2 square:1 formed:1 efficiently:3 yield:2 identify:1 handwritten:1 produced:1 knot:1 emphasizes:1 stroke:1 sharing:1 aligns:1 naturally:1 mi:1 degeneracy:1 sampled:3 gain:1 dimensionality:19 infers:1 improves:1 schedule:1 appears:2 attained:1 higher:1 tipping:1 improved:1 leen:1 done:2 furthermore:1 just:1 langford:1 working:1 nonlinear:4 defines:1 effect:2 requiring:1 true:1 isomap:8 counterpart:1 hence:3 regularization:1 inspiration:1 q0:1 self:1 generalized:7 whye:1 leftmost:1 complete:1 performs:1 gh:1 silva:1 image:9 variational:1 invoked:1 novel:1 recently:2 fi:2 patching:1 rotation:1 volume:2 interpretation:1 relating:2 expressing:1 slant:1 automatic:5 rd:1 analyzer:18 align:4 aligning:1 posterior:3 closest:2 recent:2 verlag:1 captured:1 minimum:4 additional:1 fortunately:1 maximize:1 multiple:1 full:1 infer:3 smooth:1 technical:1 calculation:1 believed:1 post:3 verbeek:1 basic:1 regression:1 noiseless:1 sometimes:1 achieved:1 receive:1 eigensolver:3 annealing:1 unlike:1 comment:1 subject:1 december:2 leveraging:1 spirit:1 call:1 easy:1 variety:1 fit:2 restrict:1 topology:1 reduce:1 idea:2 expression:2 handled:1 effort:1 penalty:1 suffer:1 interpolates:1 eigenvectors:8 transforms:1 nonparametric:2 locally:8 tenenbaum:1 notice:3 coordinating:1 write:1 group:3 key:2 neither:1 ht:1 merely:1 convert:1 run:1 fourth:1 powerful:2 throughout:1 submodels:1 yann:1 interpolants:1 trefoil:3 scaling:1 quadratic:2 topological:1 constraint:8 worked:1 ywteh:1 fukunaga:1 bdc:1 department:2 according:1 instantiates:1 belonging:1 describes:4 across:2 em:2 sam:1 smaller:1 appealing:1 n4:1 making:1 b:1 modification:1 invariant:1 equation:1 agree:2 previously:2 visualization:2 describing:1 turn:1 unusual:1 operation:2 apply:2 enforce:1 original:13 assumes:1 top:3 ghahramani:1 objective:1 already:2 added:1 quantity:1 parametric:3 costly:1 separate:1 thank:1 berlin:1 evenly:1 manifold:14 assuming:1 charting:4 index:2 relationship:2 unrolled:2 unfortunately:1 trace:1 disparate:1 unknown:2 perform:2 teh:1 upper:2 curved:2 defining:1 extended:1 hinton:3 vlassis:1 overlooked:1 specified:3 coherent:2 learned:1 beyond:1 able:1 below:1 usually:1 summarize:2 reliable:1 memory:1 mfa:13 improve:1 lk:5 geometric:1 multiplication:1 determining:4 relative:1 interesting:1 geoffrey:1 degree:1 vectorized:1 consistent:3 principle:1 share:1 translation:1 row:5 free:1 bias:1 lle:12 side:2 allow:2 saul:3 face:3 sparse:1 curve:8 dimension:9 xn:1 llc:18 calculated:2 contour:1 valid:1 simplified:1 transaction:1 approximate:1 compact:1 emphasize:1 olsen:1 status:1 global:30 latent:10 learn:2 decoupling:1 obtaining:2 excellent:1 complex:2 interpolating:1 domain:1 spread:1 linearly:1 whole:2 noise:3 ose:1 lie:1 third:3 bishop:2 emphasized:2 gating:1 offset:1 essential:2 intrinsic:1 mnist:2 kr:1 conditioned:1 nk:2 suited:1 smoothly:1 applies:2 springer:1 corresponds:1 goal:1 feasible:1 change:3 uniformly:1 averaging:1 principal:4 experimental:1 brand:3 highdimensional:1 internal:11 support:1 handling:1 |
1,297 | 2,181 | Approximate Inference and
Protein-Folding
Chen Yanover and Yair Weiss
School of Computer Science and Engineering
The Hebrew University of J erusalem
91904 Jerusalem, Israel
{cheny,yweiss} @cs.huji.ac.it
Abstract
Side-chain prediction is an important subtask in the protein-folding
problem. We show that finding a minimal energy side-chain configuration is equivalent to performing inference in an undirected
graphical model. The graphical model is relatively sparse yet has
many cycles. We used this equivalence to assess the performance of
approximate inference algorithms in a real-world setting. Specifically we compared belief propagation (BP), generalized BP (GBP)
and naive mean field (MF).
In cases where exact inference was possible, max-product BP always found the global minimum of the energy (except in few cases
where it failed to converge), while other approximation algorithms
of similar complexity did not. In the full protein data set, maxproduct BP always found a lower energy configuration than the
other algorithms, including a widely used protein-folding software
(SCWRL).
1
Introduction
Inference in graphical models scales exponentially with the number of variables.
Since many real-world applications involve hundreds of variables, it has been impossible to utilize the powerful mechanism of probabilistic inference in these applications. Despite the significant progress achieved in approximate inference, some
practical questions still remain open - it is not yet known which algorithm to use
for a given problem nor is it understood what are the advantages and disadvantages of each technique. We address these questions in the context of real-world
protein-folding application - the side-chain prediction problem.
Predicting side-chain conformation given the backbone structure is a central problem in protein-folding and molecular design. It arises both in ab-initio proteinfolding (which can be divided into two sequential tasks - the generation of nativelike backbone folds and the positioning of the side-chains upon these backbones [6])
and in homology modeling schemes (where the backbone and some side-chains are
assumed to be conserved among the homologs but the configuration of the rest of
the side-chains needs to be found).
Figure 1: Cow actin binding protein (PDB code 1pne, top) and closer view of its 6
carboxyl-terminal residues (bottom-left). Given the protein backbone (black) and
amino acid sequence, native side-chain conformation (gray) is searched for. Problem
representation as a graphical model for those carboxyl-terminal residues shown in
the bottom-right figure (nodes located at COl atom positions, edges drawn in black).
In this paper, we show the equivalence between side-chain prediction and inference
in an undirected graphical model. We compare the performance of BP, generalized
BP and naive mean field on this problem as well as comparing to a widely used
protein-folding program called SCWRL.
2
The side-chain prediction problem
Proteins are chains of simpler molecules called amino acids. All amino acids have
a common structure - a central carbon atom (COl) to which a hydrogen atom,
an amino group (N H 2 ) and a carboxyl group (COOH) are bonded. In addition,
each amino acid has a chemical group called the side-chain, bound to COl. This
group distinguishes one amino acid from another and gives its distinctive properties.
Amino acids are joined end to end during protein synthesis by the formation of
peptide bonds. An amino acid unit in a protein is called a residue. The formation
of a succession of peptide bonds generate the backbone (consisting of COl and its
adjacent atoms, N and CO, of each reside), upon which the side-chains are hanged
(Figure 1).
We seek to predict the configuration of all the side-chains relative to the backbone.
The standard approach to this problem is to define an energy function and use the
configuration that achieves the global minimum of the energy as the prediction.
2.1
The energy function
We adopted the van der Waals energy function, used by SCWRL [3], which approximates the repulsive portion of Lennard-Jones 12-6 potential. For a pair of atoms,
a and b, the energy of interaction is given by:
E(a, b) = { -k2
:'0 + k~
d> Ro
Ro ~ d ~ k1Ro
k1Ro > d
Emax
(1)
where Emax = 10, kl = 0.8254 and k2 = ~~k;' d denotes the distance between
Ro is the sum of their radii. Constant radii were used for protein's
atoms (Carbon - 1.6A, Nitrogen and Oxygen - 1.3A, Sulfur - 1.7A). For two sets
of atoms, the interaction energy is a sum of the pairwise atom interactions. The
energy surface of a typical protein in the data set has dozens to thousands local
minima.
a and band
2.2
Rotamers
The configuration of a single side-chain is represented by at most 4 dihedral angles
(denoted Xl,X2,X3 and X4)' Any assignment of X angles for all the residues defines
a protein configuration. Thus the energy minimization problem is a highly nonlinear
continuous optimization problem.
It turns out, however, that side-chains have a small repertoire of energetically preferred conformations, called rotamers. Statistical analysis of those conformations in
well-determined protein structures produce a rotamer library. We used a backbone
dependent rotamer library (by Dunbrack and Kurplus, July 2001 version). Given
the coordinates of the backbone atoms, its dihedral angles ? (defined, for the ith
residue, by Ci - 1 - Ni - Ci - Ci ) and 'IjJ (defined by Ni - Ci - Ci - NHd were
calculated. The library then gives the typical rotamers for each side-chain and their
prior probabilities.
By using the library we convert the continuous optimization problem into a discrete
one. The number of discrete variables is equal to the number of residues and the
possible values each variable can take lies between 2 and 81.
2.3
Graphical model
Since we have a discrete optimization problem and the energy function is a sum of
pairwise interactions, we can transform the problem into a graphical model with
pairwise potentials. Each node corresponds to a residue, and the state of each node
represents the configuration of the side chain of that residue. Denoting by {rd an
assignment of rotamers for all the residues then:
P({ri}) =
! e - +E({r;})
Z
!e -+ L;j E(r;)+E(r;,rj)
Z
1
Z II 'lti(ri) II 'ltijh,rj)
i
(2)
i ,j
where Z is an explicit normalization factor and T is the system "temperature"
(used as free parameter). The local potential 'ltih) takes into account the prior
probability of the rotamer Pi(ri) (taken from the rotamer library) and the energy
of the interactions between that rotamer and the backbone:
\(Ii(ri) = Pi (ri)e-,j,E(ri ,backbone)
(3)
Equation 2 requires multiplying \(I ij for all pairs of residues i, j but note that equation 1 gives zero energy for atoms that are sufficiently far away. Thus we only need
to calculate the pairwise interactions for nearby residues. To define the topology of
the undirected graph, we examine all pairs of residues i, j and check whether there
exists an assignment ri, rj for which the energy is nonzero. If it exists, we connect
nodes i and j in the graph and set the potential to be:
(4)
Figure 1 shows a subgraph of the undirected graph. The graph is relatively sparse
(each node is connected to nodes that are close in 3D space) but contains many
small loops. A typical protein in the data set gives rise to a model with hundreds
of loops of size 3.
3
Experiments
When the protein was small enough we used the max-junction tree algorithm [1] to
find the most likely configuration of the variables (and hence the global minimum
of the energy function). Murphy's implementation of the JT algorithm in his BN
toolbox for Matlab was used [10].
The approximate inference algorithms we tested were loopy belief propagation (BP),
generalized BP (GBP) and naive mean field (MF).
BP is an exact and efficient local message passing algorithm for inference in singly
connected graphs [15]. Its essential idea is replacing the exponential enumeration
(either summation or maximizing) over the unobserved nodes with series of local enumerations (a process called "elimination" or "peeling"). Loopy BP, that is
applying BP to multiply connected graphical models , may not converge due to circulation of messages through the loops [12]. However, many groups have recently
reported excellent results using loopy BP as an approximate inference algorithm
[4, 11, 5]. We used an asynchronous update schedule and ran for 50 iterations or
until numerical convergence.
GBP is a class of approximate inference algorithms that trade complexity for accuracy [15]. A subset of GBP algorithms is equivalent to forming a graph from
clusters of nodes and edges in the original graph and then running ordinary BP on
the cluster graph. We used two large clusters. Both clusters contained all nodes
in the graph but each cluster contained only a subset of the edges. The first cluster contained all edges resulting from residues, for which the difference between
its indices is less than a constant k (typically, 6). All other edges were included
in the second cluster. It can be shown that the cluster graph BP messages can
be computed efficiently using the JT algorithm. Thus this approximation tries to
capture dependencies between a large number of nodes in the original graph while
maintaining computational feasibility.
The naive MF approximation tries to approximate the joint distribution in equation 2 as a product of independent marginals qi(ri) . The marginals qi(ri) can be
found by iterating:
qi(ri)
f-
a\(li(ri) exp
(L L
qj(rj) log \(Iij(ri, rj ))
JENi
rj
(5)
where a denotes a normalization constant and Ni means all nodes neighboring i.
We initialized qi(ri) to \[Ii(ri) and chose a random update ordering for the nodes.
For each protein we repeated this minimization 10 times (each time with a different
update order) and chose the local minimum that gave the lowest energy.
In addition to the approximate inference algorithms described above, we also compared the results to two approaches in use in side-chain prediction: the SCWRL and
DEE algorithms. The Side-Chain placement With a Rotamer Library (SCWRL)
algorithm is considered one of the leading algorithms for predicting side-chain conformations [3]. It uses the energy function described above (equation 1) and a
heuristic search strategy to find a minimal energy conformation in a discrete conformational space (defined using rotamer library).
Dead end elimination (DEE) is a search algorithm that tries to reduce the search
space until it becomes suitable for an exhaustive search. It is based on a simple
condition that identifies rotamers that cannot be members of the global minimum
energy conformation [2]. If enough rotamers can be eliminated, the global minimum energy conformation can be found by an exhaustive search of the remaining
rotamers.
The various inference algorithms were tested on set of 325 X-ray crystal structures
with resolution better than or equal to 2A, R factor below 20% and length up to 300
residues. One representative structure was selected from each cluster of homologous
structures (50% homology or more) . Protein structures were acquired from Protein
Data Bank site (http://www.rcsb.org/pdb).
Many proteins contain Cysteine residues which tend to form strong disulfide bonds
with each other. A standard technique in side-chain prediction (used e.g. in
SCWRL) is to first search for possible disulfide bonds and if they exist to freeze
these residues in that configuration. This essentially reduces the search space. We
repeated our experiments with and without freezing the Cysteine residues.
Side-chain to backbone interaction seems to be much severe than side-chain to sidechain interaction - the backbone is more rigid than side-chains and its structure
assumed to be known. Therefore, the parameter R was introduced into the pairwise
potential equation, as follows:
\[Io(ro
ro) - (e -,f-E(ri ,r;))*
"J ", J -
(6)
Using R > 1 assigns an increased weight for side-chain to backbone interactions
over side-chain to side-chain interactions. We repeated our experiments both with
R = 1 and R > 1. It worth mentioning that SCWRL implicitly adopts a weighting
assumption that assigns an increased weight to side-chain to backbone interactions.
4
Results
In our first set of experiments we wanted to compare approximate inference to
exact inference. In order to make exact inference possible we restricted the possible
rotamers of each residue. Out of the 81 possible states we chose a subset whose
local probability accounted for 90% of the local probability. We constrained the size
of the subset to be at least 2. The resulting graphical model retains only a small
fraction of the loops occurring in the full graphical model (about 7% of the loops
of size 3). However, it still contains many small loops, and in particular, dozens of
loops of size 3.
On these graphs we found that ordinary max-product BP always found the global
minimum of the energy function (except in few cases where it failed to converge).
80
80
70
II!
70
..
80
.!! 50
eo
.!! 50
a.
a.
~ <1l
~
"'~ 30
<1l
"'~ 30
20
I
20
10
0
{;>
I ?
"
"
10
0
,,, 01> ~ {> .?> ..," ." ."
<9 4> <P $'
{;>
"
E(Sum-product BP) - E(Max-product BP)
I.
",,, 01> ~ {>.?>..,"."."
<9 4> <p.<p
E(Mean field) - E(Max-product BP)
80
.
70
eo
100
.!! 50
a.
gOJ
~ <1l
-
98
"'~ 30
20
10
0
{;>
.
.
.
-. - "
"
," 01> ~ {> .?> ..," ." ."
<9 4> <p.*
E(SCWRL) - E(Max-product BP)
t
96
o
94
>
c
(,)
,---
OJ
";J!. 92
90
SCWRL
Sum , R=1 Sum , R>1
nn
Max. R= 1 Max. R>1
Figure 2: Sum-product BP (top-left), naive MF (top-right) and SCWRL (bottomleft) algorithms energies are always higher than or equal to max-product BP energy.
Convergence rates for the various algorithms shown in bottom-right chart.
Sum-product BP failed to find sum-JT conformation in 1% of the graphs only. In
contrast the naive MF algorithm found the global minimum conformation for 38%
of the proteins and on 17% of the runs only. The GBP algorithm gave the same
result as the ordinary BP but it converged more often (e.g. 99.6% and 98.9% for
sum-product GBP and BP, respectively).
In the second set of experiments we used the full graphical models. Since exact
inference is impossible we can only compare the relative energies found by the
different approximate inference algorithms. Results are shown in Figure 2. Note
that, when it converged, max-product BP always found a lower energy configuration
compared to the other algorithms. This finding agrees with the observation that the
max-product solution is a "neighborhood optimum" and therefore guaranteed to be
better than all other assignments in a large region around it [1 3].
We also tried decreasing T , the system "temperature", for sum-product (in the
limit of zero temperature it should approach max-product) . In 96% of the time,
using lower temperature (T = 0.3 instead of T = 1) indeed gave a lower energy
configuration. Even at this reduced temperature, however, max-product always
found a lower energy configuration.
All algorithms converged in more than 90% of the cases. However, sum-product
converged more often than max-product (Figure 2, bottom-right) . Decreasing temperature resulted in lower convergence rate for sum-product BP algorithm (e.g.
95.7% compared to 98.15% in full size graphs using disulfide bonds). It should be
mentioned that SCWRL failed to converge on a single protein in the data set.
Applying the DEE algorithm to the side-chain prediction graphical models dramatically decreased the size of the conformational search space, though, in most cases,
the resulted space was still infeasible. Moreover, max-product BP was indifferent
;::; 3
;::; 3
.
.~
~
e::.
~
e::.
~ 2
u
2
U
:::J
:::J
rn
rn
<1' 1
<1' 1
0
Xl
x2
x3
x4
Xl
X2
Xl
X4
SCWRL buried residues success rates
Xl
X2
X3
X4
85.9% 62.2% 40.3%
25.5%
Figure 3: Inference results - success rate. SCWRL buried residues success rate
subtracted from sum-product BP (light gray), max-product BP (dark gray) and
MF (black) rates when equally weighting side-chain to backbone and side-chain to
side-chain clashes (left) and assigning increased weight for side-chain to backbone
clashes (right).
to that space reduction - it failed to converge for the same models and, when
converged, found the same conformation.
4.1
Success rate
In comparing the performance of the algorithms, we have focused on the energy of
the found configuration since this is the quantity the algorithms seek to optimize.
A more realistic performance measure is: how well do the algorithms predict the
native structure of the protein?
The dihedral angle Xi is deemed correct when it is within 40? of the native (crystal)
structure and Xl to Xi-l are correct. Success rate is defined as the portion of
correctly predicted dihedral angles.
The success rates of the conformations, inferred by both max- and sum-product
outperformed SCWRL's (Figure 3). For buried residues (residues with relative
accessibility lower than 30% [9]) both algorithms added 1% to SCWRL's Xl success
rate. Increasing the weight of side-chain to backbone interactions over side-chain
to side-chain interactions resulted in better success rates (Figure 3, right). Freezing
Cysteine residues to allow the formation of disulfide bonds slightly increased the
success rate.
5
Discussion
Recent years have shown much progress in approximate inference. We believe that
the comparison of different approximate inference algorithms is best done in the
context of a real-world problem. In this paper we have shown that for a realworld problem with many loops, the performance of belief propagation is excellent.
In problems where exact inference was possible max-product BP always found the
global minimum of the energy function and in the full protein data set, max-product
BP always found a lower energy configuration compared to the other algorithms
tested.
SCWRL is considered one of the leading algorithms for modeling side-chain conformations. However, in the last couple of years several groups reported better results
due to more accurate energy function [7], better searching algorithm [8] , or extended
rotamer library [14].
As shown, by using inference algorithms we achieved low energy conformations,
compared to existing algorithms. However, this leads only to a modest increase in
prediction accuracy. Using an energy function, which gives a better approximation
to the "true" physical energy (and particularly, assigns lowest energy to the native
structure) should significantly improve the success rate. A promising direction for
future research is to try and learn the energy function from examples. Inference
algorithms such as BP may play an important role in the learning procedure.
References
[1] R. Cowell. Introduction to inference in Bayesian networks. In Michael I. Jordan,
editor, Learning in Graphical Models. Morgan Kauffmann , 1998.
[2] Johan Desmet , Marc De Maeyer, Bart Hazes, and Ignace Lasters . The dead-end
elmination theorem and its use in protein side-chain positioning. Nature, 356:539542, 1992.
[3] Roland L. Dunbrack, Jr. and Martin Kurplus. Back-bone dependent rotamer library
for proteins: Application to side-chain predicrtion. J. Mol. Biol, 230:543- 574, 1993.
See also http://www.fccc.edu/research/labs/dunbrack/scwrlj.
[4] William T. Freeman and Egon C. Pasztor. Learning to estimate scenes from images. In
M.S. Kearns, S.A. SoHa, and D.A. Cohn , editors, Adv. Neural Information Processing
Systems 11. MIT Press, 1999.
[5] Brendan J. Frey, Ralf Koetter, and Nemanja Petrovic. Very loopy belief propagation
for unwrapping phase images. In Adv. Neural Information Processing Systems 14.
MIT Press, 200l.
[6] Enoch S. Huang, Patrice Koehl, Michael Levitt, Rohit V. Pappu, and Jay W. Ponder.
Accuracy of side-chain prediction upon near-native protein backbones generated by
ab initio folding methods. Proteins, 33(2):204- 217, 1998.
[7] Shide Liang and Nick V. Grishin. Side-chain modeling with an optimized scoring
function. Protein Sci, 11(2):322- 331, 2002.
[8] Loren L. Looger and Homme W. HeHinga. Generalized dead-end elimination algorithms make large-scale protein side-chain structure prediction tractable: implications
for protein design and structural genomics. J Mol Biol, 307(1) :429- 445 , 200l.
[9] Joaquim Mendes, Cludio M. Soare, and Maria Armnia Carrondo. mprovement of sidechain modeling in proteins with the self-consistent mean field theory method based on
an analysis of the factors influencing prediction. Biopolym ers, 50(2):111- 131 , 1999.
[10] Kevin Murphy. The bayes net toolbox for matlab . Computing Science and Statistics,
33, 200l.
[11] Kevin P. Murphy, Yair Weiss, and Micheal I. Jordan. Loopy belief propagation for
approximate inference: an empirical study. In Proceedings of Uncertainty in AI, 1999.
[12] Judea Pearl. Probabilistic R easoning in Intellig ent Systems: Networks of Plausible
Inference. Morgan Kaufmann, 1988.
[13] Yair Weiss and William T. Freeman. On the optimality of solutions of the maxproduct belief propagation algorithm. IEEE Transactions on Information Th eory,
47(2) :723- 735, 2000.
[14] Zhexin Xiang and Barry Honig. Extending the accuracy limits of prediction for sidechain conformations. J Mol Bioi, 311(2) :421-430, 200l.
[15] Jonathan S. Yedidia, William T. Freeman, and Yair Weiss. Understanding belief
propagation and its generalization. In G. Lakemayer and B. Nebel, editors, Exploring
Artificial Intelligence in the New Millennium. Morgan Kauffmann, 2002.
| 2181 |@word version:1 seems:1 open:1 seek:2 tried:1 bn:1 soare:1 reduction:1 configuration:15 contains:2 series:1 denoting:1 existing:1 clash:2 comparing:2 yet:2 assigning:1 joaquim:1 realistic:1 numerical:1 koetter:1 wanted:1 update:3 bart:1 intelligence:1 selected:1 dunbrack:3 ith:1 node:12 org:1 simpler:1 ray:1 acquired:1 pairwise:5 indeed:1 nor:1 examine:1 terminal:2 freeman:3 decreasing:2 enumeration:2 increasing:1 becomes:1 moreover:1 lowest:2 israel:1 what:1 backbone:19 finding:2 unobserved:1 ro:5 k2:2 unit:1 influencing:1 engineering:1 understood:1 local:7 limit:2 io:1 frey:1 despite:1 black:3 chose:3 equivalence:2 co:1 mentioning:1 practical:1 pappu:1 x3:3 procedure:1 empirical:1 significantly:1 pdb:2 protein:34 cannot:1 close:1 context:2 impossible:2 applying:2 www:2 equivalent:2 optimize:1 maximizing:1 jerusalem:1 focused:1 resolution:1 assigns:3 emax:2 his:1 ralf:1 searching:1 coordinate:1 kauffmann:2 play:1 exact:6 us:1 rotamers:8 cysteine:3 particularly:1 located:1 native:5 bottom:4 role:1 capture:1 thousand:1 calculate:1 region:1 cycle:1 connected:3 adv:2 ordering:1 trade:1 ran:1 mentioned:1 subtask:1 complexity:2 upon:3 distinctive:1 egon:1 joint:1 represented:1 various:2 looger:1 artificial:1 formation:3 neighborhood:1 kevin:2 exhaustive:2 whose:1 heuristic:1 widely:2 plausible:1 statistic:1 transform:1 patrice:1 advantage:1 sequence:1 net:1 interaction:13 product:25 neighboring:1 loop:8 subgraph:1 ent:1 convergence:3 cluster:9 optimum:1 extending:1 produce:1 ac:1 ij:1 school:1 conformation:15 progress:2 strong:1 c:1 predicted:1 direction:1 radius:2 correct:2 elimination:3 generalization:1 yweiss:1 repertoire:1 summation:1 exploring:1 initio:2 sufficiently:1 considered:2 around:1 exp:1 predict:2 achieves:1 nebel:1 outperformed:1 bond:6 peptide:2 agrees:1 minimization:2 mit:2 always:8 maria:1 check:1 contrast:1 brendan:1 inference:28 dependent:2 rigid:1 nn:1 typically:1 buried:3 among:1 denoted:1 constrained:1 field:5 equal:3 atom:10 eliminated:1 x4:4 represents:1 jones:1 future:1 few:2 distinguishes:1 resulted:3 murphy:3 phase:1 consisting:1 william:3 ab:2 message:3 highly:1 multiply:1 severe:1 indifferent:1 light:1 dee:3 chain:43 implication:1 accurate:1 edge:5 closer:1 modest:1 tree:1 initialized:1 minimal:2 increased:4 modeling:4 disadvantage:1 retains:1 assignment:4 loopy:5 ordinary:3 subset:4 hundred:2 reported:2 connect:1 dependency:1 petrovic:1 loren:1 huji:1 probabilistic:2 michael:2 synthesis:1 central:2 dihedral:4 huang:1 dead:3 leading:2 li:1 account:1 potential:5 de:1 view:1 try:4 bone:1 lab:1 portion:2 bayes:1 ass:1 chart:1 ni:3 accuracy:4 circulation:1 acid:7 kaufmann:1 efficiently:1 succession:1 bayesian:1 multiplying:1 worth:1 converged:5 energy:37 nitrogen:1 couple:1 judea:1 schedule:1 back:1 higher:1 wei:4 done:1 though:1 until:2 replacing:1 freezing:2 nonlinear:1 cohn:1 propagation:7 defines:1 gray:3 believe:1 homologs:1 contain:1 homology:2 true:1 hence:1 chemical:1 nonzero:1 adjacent:1 during:1 self:1 generalized:4 bonded:1 crystal:2 temperature:6 oxygen:1 image:2 recently:1 common:1 physical:1 exponentially:1 approximates:1 marginals:2 significant:1 freeze:1 ai:1 rd:1 waals:1 surface:1 recent:1 success:10 der:1 scoring:1 conserved:1 morgan:3 minimum:10 eo:2 converge:5 barry:1 july:1 ii:5 full:5 rj:6 reduces:1 positioning:2 divided:1 molecular:1 equally:1 roland:1 feasibility:1 qi:4 prediction:13 essentially:1 iteration:1 normalization:2 achieved:2 folding:7 addition:2 residue:23 decreased:1 rest:1 tend:1 undirected:4 member:1 jordan:2 structural:1 near:1 enough:2 gave:3 topology:1 cow:1 reduce:1 idea:1 qj:1 whether:1 energetically:1 passing:1 matlab:2 dramatically:1 iterating:1 conformational:2 involve:1 singly:1 dark:1 band:1 reduced:1 generate:1 http:2 goj:1 exist:1 eory:1 correctly:1 discrete:4 group:6 drawn:1 lti:1 utilize:1 graph:14 fraction:1 sum:15 convert:1 year:2 run:1 angle:5 realworld:1 powerful:1 uncertainty:1 bound:1 guaranteed:1 easoning:1 fold:1 placement:1 bp:32 x2:4 software:1 ri:15 disulfide:4 scene:1 nearby:1 optimality:1 performing:1 relatively:2 martin:1 jr:1 remain:1 slightly:1 restricted:1 intellig:1 taken:1 haze:1 equation:5 ponder:1 turn:1 mechanism:1 tractable:1 end:5 adopted:1 repulsive:1 junction:1 yedidia:1 away:1 subtracted:1 yair:4 carboxyl:3 original:2 denotes:2 top:3 running:1 remaining:1 graphical:13 maintaining:1 question:2 pne:1 quantity:1 added:1 strategy:1 distance:1 sci:1 accessibility:1 homme:1 code:1 length:1 index:1 hebrew:1 liang:1 carbon:2 rise:1 design:2 implementation:1 observation:1 pasztor:1 extended:1 rn:2 rotamer:9 inferred:1 introduced:1 pair:3 kl:1 toolbox:2 optimized:1 gbp:6 nick:1 pearl:1 address:1 below:1 program:1 max:19 including:1 oj:1 belief:7 suitable:1 homologous:1 predicting:2 nemanja:1 yanover:1 scheme:1 improve:1 millennium:1 library:9 identifies:1 deemed:1 naive:6 genomics:1 prior:2 understanding:1 rohit:1 relative:3 xiang:1 generation:1 consistent:1 editor:3 bank:1 pi:2 unwrapping:1 accounted:1 last:1 free:1 asynchronous:1 actin:1 infeasible:1 side:42 allow:1 sparse:2 van:1 calculated:1 world:4 reside:1 adopts:1 sulfur:1 far:1 transaction:1 approximate:13 preferred:1 implicitly:1 global:8 nhd:1 assumed:2 xi:2 hydrogen:1 continuous:2 search:8 promising:1 learn:1 johan:1 molecule:1 nature:1 mol:3 excellent:2 marc:1 did:1 repeated:3 amino:8 levitt:1 site:1 representative:1 jeni:1 iij:1 position:1 explicit:1 exponential:1 col:4 xl:7 lie:1 weighting:2 jay:1 peeling:1 dozen:2 theorem:1 jt:3 er:1 exists:2 essential:1 sequential:1 ci:5 occurring:1 chen:1 mf:6 soha:1 likely:1 sidechain:3 ijj:1 forming:1 failed:5 contained:3 joined:1 binding:1 cowell:1 corresponds:1 bioi:1 included:1 specifically:1 except:2 typical:3 determined:1 kearns:1 called:6 maxproduct:2 searched:1 arises:1 jonathan:1 tested:3 biol:2 mendes:1 |
1,298 | 2,182 | Recovering Articulated Model Topology from Observed
Rigid Motion
Leonid Taycher, John W. Fisher III, and Trevor Darrell
Artificial Intelligence Laboratory
Massachusetts Institute of Technology
Cambridge, MA, 02139
{lodrion, fisher, trevor}@ai.mit.edu
Abstract
Accurate representation of articulated motion is a challenging problem
for machine perception. Several successful tracking algorithms have
been developed that model human body as an articulated tree. We propose a learning-based method for creating such articulated models from
observations of multiple rigid motions. This paper is concerned with
recovering topology of the articulated model, when the rigid motion of
constituent segments is known. Our approach is based on finding the
Maximum Likelihood tree shaped factorization of the joint probability
density function (PDF) of rigid segment motions. The topology of graphical model formed from this factorization corresponds to topology of the
underlying articulated body. We demonstrate the performance of our algorithm on both synthetic and real motion capture data.
1 Introduction
Tracking human motion is an integral part of many proposed human-computer interfaces,
surveillance and identification systems, as well as animation and virtual reality systems. A
common approach to this task is to model the body as a kinematic tree, and reformulate
the problem as articulated body tracking[6]. Most of the state-of-the-art systems rely on
predefined kinematic models [16]. Some methods require manual initialization, while other
use heuristics [12], or predefined protocols [10] to adapt the model to observations.
We are interested in a principled way to recover articulated models from observations. The
recovered models may then be used for further tracking and/or recognition. We would
like to approach model estimation as a multistage problem. In the first stage the rigidly
moving segments are tracked independently; at the second stage, the topology of the body
(the connectivity between the segments) is recovered. After the topology is determined, the
joint parameters may be determined.
In this paper we concentrate on the second stage of this task, estimating the underlying
topology of the observed articulated body, when the motion of the constituent rigid bodies
is known. We approach this as a learning problem, in the spirit of [17]. If we assume that
the body may be modeled as a kinematic tree, and motion of a particular rigid segment is
known, then the motions of the rigid segments that are connected through that segment are
independent of each other. That is, we can model a probability distribution of the full body-
pose as a tree-structured graphical model, where each node corresponds to pose of a rigid
segment. This observation allows us to formulate the problem of recovering topology of
an articulated body as finding the tree-shaped graphical model that best (in the Maximum
Likelihood sense) describes the observations.
2 Prior Work
While state-of-the-art tracking algorithms [16] do not address either model creation or
model initialization, the necessity of automating these two steps has been long recognized.
The approach in [10] required a subject to follow a set of predefined movements, and
recovered the descriptions of body parts and body topology from deformations of apparent
contours. Various heuristics were used in [12] to adapt an articulated model of known
topology to 3D observations. Analysis of magnetic motion capture data was used by [14]
to recover limb lengths and joint locations for known topology, it also suggested similar
analysis for topology extraction. A learning based approach for decomposing a set of
observed marker positions and velocities into sets corresponding to various body parts was
described in [17]. Our work builds on the latter two approaches in estimating the topology
of the articulated tree model underlying the observed motion.
Several methods have been used to recover multiple rigid motions from video, such as
factorization [3, 18], RANSAC [7], and learning based methods [9]. In this work we assume that the 3-D rigid motions has been recovered and are represented using a 2-D Scaled
Prismatic Model (SPM).
3 Representing Pose and Motion
A 2-D Scaled Prismatic Model (SPM) was proposed by [15] and is useful for representing
image motion of projections of elongated 3-D objects. It is obtained by orthographically
?projecting? the major axis of the object to the image plane. The SPM has four degrees of
freedom: in-plane translation, rotation, and uniform scale. 3-D rigid motion of an object,
may be simulated by SPM transformations, using in-plane translation for rigid translation,
and rotation and uniform scaling for plane-parallel and out-of-plane rotations respectively.
SPM motion (or pose) may be expressed as a linear transformation in projective space as
!
a ?b e
M= b a f
(1)
0 0 1
Following [13] we have chosen to use exponential coordinates, derived from constant velocity equations, to parameterize motion.
An SPM transformation may be represented as an exponential map
?
M = e?
c
?? = ? ?
0
??
c
0
vx
vy
0
!
? ?
vx
? vy ?
? = ?? ?
?
c
(2)
In this representation vx is a horizontal velocity, vy ? vertical velocity, ? ? angular velocity,
and c is a rate of scale change. ? is analogous to time parameter. Note that there is an
inherent scale ambiguity, since ? and (vx , vy , ?, c)T may be chosen arbitrarily, as long as
?
e? = M.
It can be shown ([13]) that if the SPM transformation is a combination of scaling and
rotation, it may be expressed by the sum of two twists, with coincident centers (u x , uy )T
of rotation and expansion.
?
?
? ?
uy
?ux
?c
??ux ?
??uy ? ???
? = ??
+c?
=
1 ?
0 ? ?
0
1
?
?c
?
?
ux
? ? uy ?
?? ? ?
1
c
??
1
(3)
While ?pure? translation, rotation or scale have intuitive representation with twists, the
combination or rotation and scale does not. We propose a scaled twist representation, that
preserves the intuitiveness of representation for all possible SPM motions. We want to
separate the ?direction? of motion (the direction of translation or the relative amounts of
rotation and scale) from the amount of motion.
If the transformation involves rotation and/or scale, then we choose ? so that ||(?, c)|| 2 = 1,
and then use eq. 3 to compute the center of rotation/expansion. The computation may be
expressed as a linear transformation:
??
?
?
?ux ? ?
? ? ?
? = ? uy ? = ?
??? ?
?
?
c
?
?
?
?
? 2 + c?2
c
?
? ?? 2 +?
c2
?
?
?
? 2 +?
c2
?
? ?? 2?+?
c2
c
?
? ?? 2 +?
c2
?
1
?
? 2 +?
c2
?
1
?
? 2 +?
c2
where ? = (?
vx , v?y , ?
? , c?)T .
? ?
? 1
? ?v?x ?
?? ?
? ?v?y ?
?? ?
? ?
?
?
c?
(4)
The the pure translational motion (? = c = 0) may be regarded as an infinitely small
rotation about a point at infinity, e.g. the translation by l in the direction (u x , uy ) may be
?u
represented as ? = lim??0 (l|?|, ? y , u?x , ?, 0)T , but we choose a direct representation
?p 2
?
v?x + v?y2
?
?
?ux ? ?
? ? ?
? = ?uy ? = ?
?0? ?
?
0
?
In both cases ? = A(1, ??T )T , and
?3
?
det(A) =
??1
? 21 2
v
?x +?
vy
?
?
1
2 +?
2
v
?x
vy
1
1
? ?
? 1
? ?v?x ?
?? ?
? ?v?y ?
???
?
? ?
c?
?=
6 0 ? c 6= 0 (rotation/scaling)
? = 0 ? c = 0 (pure translation)
(5)
(6)
Note that ?I = (0, ux , uy , ?, c)T represents identity transformation for any ux , uy , ?, and
c. It is always reported as ?I = 0.
4 Learning Articulated Topology
We wish to infer the underlying topology of an articulated body from noisy observations
of a set of rigid body motions. Towards that end we will adopt a statistical framework for
fitting a joint probability density. As a practical matter, one must make choices regarding
density models; we discuss one such choice although other choices are also suitable.
We denote the set of observed motions of N rigid bodies at time t, 1 ? t ? F as a set
{Mts |1 ? s ? N }. Graphical models provide a useful methodology for expressing the dependency structure of a set of random variables (cf. [8]). Variables M i with observations
{Mti |1 ? t ? F } are assigned to the vertices of a graph, while edges between nodes indicate dependency. We shall denote presence or absence of an edge between two variables,
Mi and Mj by an index variable Eij , equal to one if an edge is present and zero otherwise.
Furthermore, if the corresponding graphical model is a spanning tree, it can be expressed
as a product of conditional densities (e.g. see [11])
Y
PM (M1 , . . . , MN ) =
PMs |pa(Ms ) (Ms |pa (Ms ))
(7)
Ms
where pa(Ms ) is the parent of Ms . While multiple nodes may have the same parent, each
individual node has only one parent node. Furthermore, in any decomposition one node
(the root node) has no parent. Any node (variable) in the model can serve as the root node
[8]. Consequently, a tree model constrains E. Of the possible tree models (choices of
E), we wish to choose the maximum likelihood tree which is equivalent to the minimum
entropy tree [4]. The entropy of a tree model can be written
X
X
H(M ) =
H(Ms ) ?
I(Mi ; Mj )
(8)
s
Eij =1
where H(Ms ) is the marginal entropy of each variable and I(Mi ; Mj ) is the mutual information between nodes Mi and Mj and quantifies their statistical dependence. Consequently, the minimum entropy tree corresponds to the choice of E which minimizes the
sum of the pairwise mutual informations [1]. The tree denoted by E can be found via the
maximum spanning tree algorithm [2] using I(Mi ; Mj ) for all i, j as the edge weights.
Our conjecture is that if our data are sampled from a variety of motions the topology of
the estimated density model is likely to be the same as the topology of the articulated body
model. It follows from the intuition that when considering only pairwise relationships, the
relative motions of physically connected bodies will be most strongly related.
4.1 Estimation of Mutual Information
Computing the minimum entropy spanning tree requires estimating the pairwise mutual
informations between rigid motions Mi and Mj for all i, j pairs. In order to do so we
must make a choice regarding the parameterization of motion and a probability density
over that parameterization; to estimate articulated topology it is sufficient to use the the
Scaled Prismatic Model with twist parameterization described in Section 3).
4.2 Estimating Motion Entropy
We parameterize rigid motion, Mti , by the vector of quantities ?it (cf. Eq. 2). In general,
H(Mi ) 6= H(?i ),
(9)
but since there is a one-to-one correspondence between the M i ?s and ?i ?s [4], we can
estimate the I(Mi ; Mj ) by first computing ?it , ?jt from Mti , Mtj
I(Mi ; Mj ) = I(?i ; ?j ) = H(?j ) ? H(?j |?i )
Furthermore, if the relative motion Mj|i between segments si and
assumed to be independent of Mi , it can be shown that
sj (Mjt
(10)
=
H(?j |?i ) = H(log Mi Mj|i | log Mi ) = H(log Mj|i ) = H(?j|i ).
t
Mit Mj|i
)
is
(11)
We wish to use scaled twists (Section 3) to compute the entropies involved. Since the involved quantities are in the linear relationship ? = A(1, ??T )T (Eqs. 4 and 5), the entropies
are related,
H(?) = H(? ) ? E[log det(A)],
(12)
where E[log det(A)] may be estimated using Equation 6.
4.3 Estimating the Motion Kernel
In order to estimate the entropy of motion, we need to estimate the probability density
based on the available samples. Since the functional form of the underlying density is not
known we have chosen to use kernel-based density estimator,
X
p?(? ) = ?
K(? ; ?i ).
(13)
i
Since our task is to determine the articulated topology, we wish to concentrate on ?spatial?
features of the transformation, center of rotation for rotational motion, and the direction of
translation for translational, that correspond to two common kinds of joints, spherical and
prismatic. Thus we need to define a kernel function K(?1 ; ?2 ) that captures the following
notion of ?distance? between the motions:
1. If ?1 and ?2 do not represent pure translational motions, then they should be considered to be close if their centers of rotation are close.
2. If ?1 and ?2 are pure translations, then they should be considered close if their
directions are close.
3. If ?1 and ?2 represent different types of motion (i.e. rotation/scale vs. translation),
then they are arbitrarily far apart.
4. The identity transformation (? = 0) is equidistant from all possible transformations (since any (ux , uy , ?, c)T combined with ? = 0 produces identity)
One kernel that satisfies these requirements is the following:
?
KR ((ux1 , uy1 ); (ux2 , uy2 ))
?
?
?
?
?
?
?
?
?
KT ((ux1 , uy1 ); (ux2 , uy2 ))
?
?
?
?
?
?
?
?
0
K(?1 ; ?2 ) =
?
?
?
?
?
?0
?
?
?
?
?
?
?
?
?(0)
?
?
?
Condition 1
(?1 6= 0 ? c1 6= 0) ? (?2 6= 0 ? c2 6= 0)
Condition 2
?1 = 0 ? c 1 = 0 ? ? 2 = 0 ? c 2 = 0
Condition 3
(?1 6= 0 ? c1 6= 0) ? (?2 = 0 ? c2 = 0)
Condition 3
(?1 = 0 ? c1 = 0) ? (?2 6= 0 ? c2 6= 0)
Condition 4.
?1 = 0 ? ? 2 = 0
(14)
where KR and KT are Gaussian kernels with covariances estimated using methods from
[5].
5 Implementation
The input to our algorithm is a set of SPM poses (Section 3) {Pts |1 ? s ? S, 1 ? t ? T },
where S is the number of tracked rigid segments and F is the number of frames. In order
to compute the mutual information between the motion of segments s 1 and s2 , we first
compute motions of segment s1 in frames 1 < t ? F relative to its position in frame
t1 = 1,
Mts11t = Pts1 (Pts11 )?1 ,
(15)
?1
and the transformation of s2 relative to s1 (with the relative pose Ps2 |s1 = (Ps1 ) Ps2 ),
Mts12t|s1 = ((Pts1 )?1 Pts2 )((Pts11 )?1 .Pts12 )?1
(16)
t
The parameter vectors ?st21 t and ?st21|s
are then extracted from the transformation matrices
1
Ms2 and Ms2 |s1 (cf. Section 3), and the mutual information is estimated as described in
Section 4.2.
6 Results
We have tested our algorithm both on synthetic and motion capture data. Two synthetic sequences were generated with the following steps. First, the rigid segments were positioned
by randomly perturbing parameters of the corresponding kinematic tree structure. A set of
feature points was then selected for each segment. At each time step point positions were
computed based on the corresponding segment pose, and perturbed with Gaussian noise
with zero mean and standard deviation of 1 pixel. The inputs to the algorithm were the segment poses re-estimated from the feature point coordinates. In the motion capture-based
experiment, the segment poses were estimated from the marker positions.
The results of the experiments are shown in the Figures 6.1, 6.2 and 6.3. The first experiment involved a simple kinematic chain with 3 segments in order to demonstrate the
operation of the algorithm. The system has a rotational joint between S 1 and S2 and prismatic joint between S2 and S3 .
The sample configurations of the articulated body are shown in the first row of the Figures
6.1. The graph computed using method from Section 4.2 and the corresponding maximum
spanning tree are in Figures 6.1(d, e).
The second experiment involved a humanoid torso-like synthetic model containing 5 rigid
segments. It was processed in a way similar to the first experiment. The results are shown
in Figure 6.2.
For the human motion experiment, we have used motion capture data of a dance sequence
(Figure 6.3(a-c)). The rigid segment motion was extracted from the positions of the markers
tracked across 220 frames (the marker correspondence to the body locations was known).
The algorithm was able to correctly recover the articulated body topology (Compare Figures 6.3(e) and 6.3(a)), when provided only with the extracted segment poses. The dance is
a highly structured activity, so not all degrees of freedom were explored in this sequence,
and mutual information between some unconnected segments (e.g. thighs S 3 and S7 ) was
determined to be relatively large, although this did not impact the final result.
7 Conclusions
We have presented a novel general technique for recovering the underlying articulated
structure from information about rigid segment motion. Our method relies on only a very
weak assumption, that this structure may be represented by a tree with unknown topology.
While the results presented in this paper were obtained using the Scaled Prismatic Model
and non-parametric density estimator, our methodology does not rely on either modeling
assumption.
References
[1] C. K. Chow and C. N. Liu. Approximating discrete probability distributions with dependence
trees. IEEE Transactions on Information Theory, IT-14(3):462?467, May 1968.
[2] Thomas H. Cormen, Charles E. Leiserson, and Ronald L. Rivern. Introduction to Algorithms.
MIT Press, Cambridge, MA, 1990.
[3] Joao Paolo Costeira and Takeo Kanade. A multibody factorization method for independently
moving objects. International Journal of Computer Vision, 29(3):159?179, 1998.
[4] T. M. Cover and J. A. Thomas. Elements of Information Theory. John Wiley & Sons, Inc., New
York, 1991.
[5] Luc Devroye. A Course in Density Estimation, volume 14 of Progress in Probability and Statistics. Birkhauser, Boston, 1987.
S3
S2
S1
(a)
(b)
S1
S2
(c)
S3
S1
S1
S2
S2
S3
S3
(d)
(e)
Figure 6.1: Simple kinematic chain topology recovery. The first row shows 3 sample frames
from a 100 frame synthetic sequence. The adjacency matrix of the mutual information
graph is shown in (d), with intensities corresponding to edge weights. The vertices in
the graph correspond to the rigid segments labeled in (a). (e) is the recovered articulated
topology.
S4
S2
S3
S5
S1
(b)
(a)
S1
S2
S3
S4
(c)
S5
S1
S2
S3
S4
S5
(d)
(e)
Figure 6.2: Humanoid torso synthetic test. The sample frames from a randomly generated
150 frame sequence are shown in (a), (b), and (c). The adjacency matrix of the mutual
information graph is shown in (d), with intensities corresponding to edge weights. The
vertices in the graph correspond to the rigid segments labeled in (a). (e) is the recovered
articulated topology.
S1
S9
S5
S8
S4
S7
S3
S6
S2
(a)
(b)
S1
S2
S3
S4
S5
S6
S7
(c)
S8
S9
S1
S2
S3
S4
S5
S6
S7
S8
(d)
S9
(e)
Figure 6.3: Motion Capture based test. (a), (b), and (c) are the sample frames from a 220 frame
sequence. The adjacency matrix of the mutual information graph is shown in (d), with intensities
corresponding to edge weights. The vertices in the graph correspond to the rigid segments labeled in
(a). (e) is the recovered articulated topology.
[6] David C. Hogg. Model-based vision: A program to see a walking person. Image and Vision
Computing, 1(1):5?20, 1983.
[7] Yi-Ping Hung, Cheng-Yuan Tang, Sheng-Wen Shin, Zen Chen, and Wei-Song Lin. A 3d featurebased tracker for tracking multiple moving objects with a controlled binocular head. Technical
report, Academia Sinica Institute of Information Science, 1995.
[8] Finn Jensen. An Introduction to Bayesian Networks. Springer, 1996.
[9] N. Jojic and B.J. Frey. Learning flexible sprites in video layers. In Computer Vision and Pattern
Recognition, pages I:199?206, 2001.
[10] Ioannis A. Kakadiaris and Dimirti Metaxas. 3d human body acquisition from multiple views.
In Proc. Fifth International Conference on Computer Vision, pages 618?623, 1995.
[11] Marina Meila. Learning Mixtures of Trees. PhD thesis, MIT, 1998.
[12] Ivana Mikic, Mohan Triverdi, Edward Hunter, and Pamela Cosman. Articulated body posture
estimation from multi-camera voxel data. In Computer Vision and Pattern Recognition, 2001.
[13] Richard M. Murray, Zexiang Li, and S. Shankar Sastry. A Mathematical Introduction to Robotic
Manipulation. CRC Press, 1994.
[14] J. O?Brien, R. E. Bodenheimer, G. Brostow, and J. K. Hodgins. Automatic joint parameter
estimation from magnetic motion capture data. In Graphics Interface?2000, pages 53?60, 2000.
[15] James M. Regh and Daniel D. Morris. Singularities in articulated object tracking with 2-d and
3-d models. Technical report, Digital Equipment Corporation, 1997.
[16] Hedvig Sidenbladh, Michael J. Black, and David J. Fleet. Stochastic tracking of 3d human
figures using 2d image motion. In Proc. European Conference on Computer Vision, 2000.
[17] Yang Song, Luis Goncalves, Enrico Di Bernardo, and Pietro Perona. Monocular perception
of biological motion - detection and labeling. In Proc. International Conference on Computer
Vision, pages 805?812, 1999.
[18] Ying Wu, Zhengyou Zhang, Thomas S. Huang, and John Y. Lin. Multibody grouping via
orthogonal subspace decomposition. In Proc. IEEE Conf. on Computer Vision and Pattern
Recognition, 2001.
| 2182 |@word decomposition:2 covariance:1 necessity:1 configuration:1 liu:1 daniel:1 brien:1 recovered:7 si:1 must:2 luis:1 written:1 john:3 ronald:1 takeo:1 academia:1 v:1 intelligence:1 selected:1 parameterization:3 plane:5 node:10 location:2 mtj:1 zhang:1 mathematical:1 c2:9 direct:1 brostow:1 yuan:1 fitting:1 pairwise:3 multi:1 spherical:1 considering:1 provided:1 estimating:5 underlying:6 joao:1 multibody:2 kind:1 minimizes:1 developed:1 ps2:2 finding:2 transformation:12 corporation:1 bernardo:1 scaled:6 t1:1 frey:1 rigidly:1 black:1 initialization:2 challenging:1 factorization:4 projective:1 uy:10 practical:1 camera:1 shin:1 projection:1 close:4 s9:3 shankar:1 equivalent:1 elongated:1 map:1 center:4 independently:2 formulate:1 recovery:1 pure:5 estimator:2 regarded:1 s6:3 notion:1 coordinate:2 analogous:1 pt:1 thigh:1 pa:3 velocity:5 element:1 recognition:4 intuitiveness:1 walking:1 labeled:3 featurebased:1 observed:5 capture:8 parameterize:2 connected:2 movement:1 principled:1 intuition:1 constrains:1 multistage:1 segment:26 creation:1 serve:1 joint:8 various:2 represented:4 articulated:25 artificial:1 labeling:1 apparent:1 heuristic:2 tested:1 otherwise:1 statistic:1 noisy:1 final:1 sequence:6 propose:2 product:1 intuitive:1 description:1 constituent:2 parent:4 darrell:1 requirement:1 produce:1 object:6 pose:10 progress:1 eq:3 edward:1 recovering:4 involves:1 indicate:1 concentrate:2 direction:5 stochastic:1 human:6 vx:5 virtual:1 adjacency:3 crc:1 require:1 biological:1 singularity:1 tracker:1 considered:2 major:1 adopt:1 estimation:5 proc:4 cosman:1 mit:4 always:1 gaussian:2 surveillance:1 derived:1 likelihood:3 equipment:1 sense:1 rigid:24 chow:1 perona:1 interested:1 pixel:1 translational:3 flexible:1 denoted:1 art:2 spatial:1 mutual:10 marginal:1 equal:1 shaped:2 extraction:1 represents:1 report:2 hedvig:1 inherent:1 richard:1 wen:1 randomly:2 preserve:1 individual:1 freedom:2 detection:1 highly:1 kinematic:6 leiserson:1 mixture:1 chain:2 predefined:3 accurate:1 kt:2 integral:1 edge:7 orthogonal:1 tree:22 re:1 deformation:1 modeling:1 cover:1 vertex:4 deviation:1 uniform:2 successful:1 graphic:1 reported:1 dependency:2 perturbed:1 synthetic:6 combined:1 person:1 density:11 international:3 automating:1 michael:1 connectivity:1 thesis:1 ambiguity:1 containing:1 choose:3 zen:1 huang:1 mjt:1 conf:1 creating:1 li:1 ioannis:1 matter:1 inc:1 root:2 view:1 recover:4 parallel:1 formed:1 correspond:4 weak:1 identification:1 bayesian:1 metaxas:1 zhengyou:1 hunter:1 ping:1 manual:1 trevor:2 acquisition:1 involved:4 james:1 mi:12 di:1 sampled:1 massachusetts:1 lim:1 torso:2 positioned:1 follow:1 methodology:2 costeira:1 wei:1 strongly:1 furthermore:3 angular:1 stage:3 binocular:1 sheng:1 horizontal:1 marker:4 spm:9 y2:1 assigned:1 jojic:1 laboratory:1 m:8 pdf:1 demonstrate:2 motion:50 interface:2 image:4 novel:1 charles:1 common:2 rotation:15 functional:1 mt:1 tracked:3 twist:5 perturbing:1 volume:1 s8:3 m1:1 expressing:1 s5:6 cambridge:2 ai:1 automatic:1 meila:1 sastry:1 pm:2 hogg:1 moving:3 apart:1 manipulation:1 arbitrarily:2 yi:1 minimum:3 recognized:1 determine:1 multiple:5 full:1 infer:1 technical:2 adapt:2 long:2 lin:2 marina:1 controlled:1 impact:1 ransac:1 vision:9 physically:1 represent:2 kernel:5 c1:3 want:1 enrico:1 subject:1 spirit:1 presence:1 yang:1 iii:1 concerned:1 ps1:1 variety:1 equidistant:1 topology:25 regarding:2 det:3 fleet:1 orthographically:1 s7:4 song:2 sprite:1 york:1 useful:2 amount:2 s4:6 morris:1 processed:1 vy:6 s3:11 estimated:6 correctly:1 discrete:1 shall:1 paolo:1 four:1 graph:8 mti:3 pietro:1 prismatic:6 sum:2 taycher:1 ms2:2 wu:1 scaling:3 layer:1 correspondence:2 cheng:1 activity:1 infinity:1 relatively:1 conjecture:1 structured:2 combination:2 cormen:1 describes:1 across:1 son:1 s1:15 projecting:1 equation:2 monocular:1 discus:1 finn:1 end:1 available:1 decomposing:1 operation:1 limb:1 magnetic:2 thomas:3 cf:3 graphical:5 build:1 murray:1 approximating:1 quantity:2 posture:1 parametric:1 dependence:2 subspace:1 distance:1 separate:1 sidenbladh:1 simulated:1 spanning:4 length:1 devroye:1 modeled:1 index:1 reformulate:1 relationship:2 rotational:2 ying:1 sinica:1 implementation:1 unknown:1 vertical:1 observation:8 coincident:1 unconnected:1 head:1 frame:10 intensity:3 david:2 pair:1 required:1 address:1 able:1 suggested:1 perception:2 pattern:3 program:1 video:2 suitable:1 rely:2 mn:1 representing:2 technology:1 axis:1 prior:1 relative:6 goncalves:1 digital:1 humanoid:2 degree:2 sufficient:1 translation:10 row:2 course:1 institute:2 fifth:1 contour:1 far:1 voxel:1 transaction:1 sj:1 robotic:1 assumed:1 quantifies:1 reality:1 kanade:1 mj:12 expansion:2 european:1 protocol:1 did:1 hodgins:1 s2:14 noise:1 animation:1 body:23 wiley:1 position:5 wish:4 exponential:2 tang:1 jt:1 jensen:1 explored:1 grouping:1 kr:2 phd:1 mohan:1 chen:1 boston:1 pamela:1 entropy:9 eij:2 likely:1 infinitely:1 expressed:4 tracking:8 ux:8 springer:1 corresponds:3 satisfies:1 relies:1 extracted:3 ma:2 conditional:1 identity:3 consequently:2 towards:1 luc:1 leonid:1 fisher:2 change:1 absence:1 determined:3 birkhauser:1 latter:1 dance:2 hung:1 |
1,299 | 2,183 | Half-Lives of EigenFlows for Spectral Clustering
Chakra Chennubhotla & Allan D. Jepson
Department of Computer Science, University of Toronto, Canada M5S 3H5
chakra,jepson @cs.toronto.edu
Abstract
Using a Markov chain perspective of spectral clustering we present an
algorithm to automatically find the number of stable clusters in a dataset.
The Markov chain?s behaviour is characterized by the spectral properties
of the matrix of transition probabilities, from which we derive eigenflows
along with their halflives. An eigenflow describes the flow of probability mass due to the Markov chain, and it is characterized by its eigenvalue, or equivalently, by the halflife of its decay as the Markov chain
is iterated. A ideal stable cluster is one with zero eigenflow and infinite half-life. The key insight in this paper is that bottlenecks between
weakly coupled clusters can be identified by computing the sensitivity
of the eigenflow?s halflife to variations in the edge weights. We propose
a novel E IGEN C UTS algorithm to perform clustering that removes these
identified bottlenecks in an iterative fashion.
1 Introduction
We consider partitioning a weighted undirected graph? corresponding to a given dataset?
into a set of discrete clusters. Ideally, the vertices (i.e. datapoints) in each cluster should
be connected with high-affinity edges, while different clusters are either not connected or
are connnected only by a few edges with low affinity. The practical problem is to identify
these tightly coupled clusters, and cut the inter-cluster edges.
Many techniques have been proposed for this problem, with some recent success being obtained through the use of spectral methods (see, for example, [2, 4, 5, 11, 12]). Here we use
the random walk formulation of [4], where the edge weights are used to construct a Markov
defines a random walk on the graph to
transition probability matrix, . This matrix
be partitioned. The eigenvalues and eigenvectors of
provide the basis for deciding on a
particular segmentation. In particular, it has been shown that for weakly coupled clusters, the leading eigenvectors of
will be roughly piecewise constant [4, 13, 5]. This
result motivates many of the current spectral clustering algorithms. For example in [5], the
number of clusters must be known a priori, and the -means algorithm is used on the
leading eigenvectors of
in an attempt to identify the appropriate piecewise constant
regions.
In this paper we investigate the form of the leading eigenvectors of the Markov matrix .
Using some simple image segmentation examples we confirm that the leading eigenvectors
of
are roughly piecewise constant for problems with well separated clusters. However,
we observe that for several segmentation problems that we might wish to solve, the coupling between the clusters is significantly stronger and, as a result, the piecewise constant
approximation breaks down.
Unlike the piecewise constant approximation, a perfectly general view is that the eigenvectors of
determine particular flows of probability along the edges in the graph. We refer
to these as eigenflows since they are characterized by their associated eigenvalue , which
specifies the flow?s overall rate of decay. Instead of measuring the decay rate in terms of
the eigenvalue , we find it more convenient to use the flow?s halflife , which is simply
defined by
. Here is the number of Markov chain steps needed to reduce
the particular eigenflow to half its initial value. Note that as approaches the half-life
approaches infinity.
From the perspective of eigenflows, a graph representing a set of weakly coupled clusters produces eigenflows between the various clusters which decay with long halflives. In
contrast, the eigenflows within each cluster decay much more rapidly. In order to identify clusters we therefore consider the eigenflows with long halflives. Given such a slowly
decaying eigenflow, we identify particular bottleneck regions in the graph which critically
restrict the flow (cf. [12]). To identify these bottlenecks we propose computing the sensitivity of the flow?s halflife with respect to perturbations in the edge weights.
We implement a simple spectral graph partitioning algorithm which is based on these ideas.
We first compute the eigenvectors for the Markov transition matrix, and select those with
long halflives. For each such eigenvector, we identify bottlenecks by computing the sensitivity of the flow?s halflife with respect to perturbations in the edge weights. In the current
algorithm, we simply select one of these eigenvectors in which a bottleneck has been identified, and cut edges within the bottleneck. The algorithm recomputes the eigenvectors and
eigenvalues for the modified graph, and continues this iterative process until no further
edges are cut.
2 From Affinities to Markov Chains
Following
the formulation in [4], we consider an undirected graph
with vertices , for
, and edges with non-negative weights . Here the weight represents the affinity of vertices and . The edge affinities are assumed to be symmetric,
that is, . A Markov chain is defined using these affinities by setting the transition
probability !" from vertex # to vertex to be proportional to the edge affinity, $ .
That is, !" %'&) (+* where &,.-./10 gives the normalizing factor which ensures
- /10 !2 23 . In matrix notation, the *affinities are represented by a symmetric 46574
* 8 , with elements $ , and the transition probability matrix 9;:!<
= is given by
matrix
>.8@? (A* B?C
Notice that the 4F524 matrix
diag :& D&
*
is not in general symmetric.
/
=E
(1)
defines the random walk of a particle on the graph
This transition probability matrix
. Suppose the initial probability of the particle being at vertex is GIH , for JK;L4 .
Then, the probability of the particle being initially at vertex and taking edge
is
!"
N MGIH . In matrix notation, the probability of the particle ending
N up anyN of the vertices
N
<O: * PQQQRM / = after one step is given by the distribution G *
G%H , where GTSU
:VGWS EXGYS = .
*
/
For analysis it is convenient to consider the matrix Z[.? (+*\ P ]? *\ P , which is similar to
(where ? is as given
N in Eq. (1)). The matrix Z therefore has the N same spectrum as
and any eigenvector ^ of Z must correspond to an eigenvector ? *\ P ^ of
with the same
eigenvalue. Note that Z_C? (+*\ P ]? *M\ P `? (A*M\ P 8@? (A* ? *\ P `? (A*M\ P 8@? (A*M\ P , and
therefore Z is a symmetric 4a524 matrix since 8 is symmetric while ? is diagonal.
The advantage of considering the matrix Z over
is that the symmetric eigenvalue problem is more stable to small perturbations, and is computationally much more tractable.
Since the matrix Z is symmetric, it has an orthogonal decomposition of the form:
Z[.bTcdbTef
(2)
(a)
(b)
(c)
(d)
(e)
Figure 1: (a-c) Three random images each having an occluder in front of a textured background. (d-e) A pair of eye images.
N N
N
bC ^ * ^ P QQQ+ ^ are the eigenvectors and c is a diagonal matrix of eigenvalQQQ+D / sorted/ in decreasing order. While the eigenvectors have unit length,
* DIP ,Lthe
^ ;
eigenvalues are real and have an absolute value bounded by 1, .
S
S
where
ues
N
The eigenvector representation provides a simple way to capture the Markovian relaxation
process [12]. For example, consider propagating the Markov chain for iterations. The
transition matrix after iterations, namely ] , can be represented as:
'? M* \ P b c b e ? (+*\ P
(3)
Therefore the probability distribution for the particle being at vertex
steps
N after
N
N of
,
G
H
@
G
G
H
the randomN walk, given
that
the
initial
probability
distribution
was
,
is
N
N
? *\ P bTcf H N , where H b e ? (+*\ P G@H provides the expansion coefficients of the initial
distribution G%H in terms of the
of Z . As
, the Markov chain approaches
N eigenvectors
N
the stationary distribution ,
e . Assuming the graph
is connected with edges
having non-zero weights, it is convenient to interpret the Markovian relaxation process as
N
N
N
perturbations to the stationary distribution, G
- / 0 P
, where * is
N
N
P N
associated with the stationary distribution and ? *M\ ^ .
3 EigenFlows
N
Let G@H be an initial probability distribution for a random particle to be at the vertices of
the graph
. By the definition of the Markov chain, recall that the probability of making
the transition from vertex to is the probability of starting in vertex , times the
conditional probability of taking edge given that the particle is at vertex , namely
! GIH . Similarly, the probability of making the transition in the reverse direction is ! D G)H .
The net flow of probability mass along edge from # to is therefore the difference
!" MGIH !UD VG)H .N It then follows that
N the net flow of probability mass from vertex to
is given by
: G,H#= , where : G%H#= is the : J= -element of the 4F5 4 matrix
K: G
N
diag : G
N
N
H = diag : G H = ]ef
N
Notice that
diag : G%H
= , and therefore
e for
). This expresses the fact that the flow from to
e
H =f
(i.e.
isisjustantisymmetric
of the flow
the opposite sign
in the reverse
direction. Furthermore, it can be shown that K: N =T for any
with _ ,
stationary distribution . Therefore, the flow is caused by the eigenvectors
N
and hence we analyze the rate of decay of these eigenflows K
: =.
(4)
For illustration purposes we begin by considering an ensemble of random test images
formed from two independent samples of 2D Gaussian filtered white noise (see Fig. 1a-c).
One sample is used to form the 5a background image, and a cropped 5
fragment
of second sample is used for the foreground region. A small constant bias is added to the
foreground region.
(a)
(b)
(c)
Figure 2: (a) Eigenmode (b) corresponding eigenflow (c) gray value at each pixel corresponds to the maximum of the absolute sensitivities of all the weights on edges connected
to a pixel (not including itself). Dark pixels indicate high absolute sensitivities.
A graph clustering problem is formed where each pixel in a test image is associated with a
vertex of the graph
. The edges in
are defined by the standard 8-neighbourhood of each
pixel (with pixels at the edges and corners of the image only having 5 and 3 neighbours,
respectively). The edge weight
between
neighbouring vertices
and is given by the
N
N
N
affinity N
: Y: = Y: =M= P $: P = , where Y: S = is the test image brightness
, where is the median
at pixel
S and is a grey-level standard deviation. We use
absolute difference of gray levels between all neighbouring pixels and ;L .
This generative process provides an ensemble of clustering problems which we feel are
representative of the structure of typical image segmentation problems. In particular, due
to the smooth variation in gray-levels, there is some variability in the affinities within both
foreground and background regions. Moreover, due to the use of independent samples for
the two regions, there is often a significant step in gray-level across the boundary between
the two regions. Finally, due to the small bias used, there is also a significant chance for
pixels on opposite sides of the boundary to have similar gray-levels, and thus high affinities.
This latter property ensures that there are some edges with significant weights between the
two clusters in the graph associated with
N the foreground and background pixels.
N
along with its eigenflow, K: S = .
In Figure 2 we plot one eigenvector, , of the matrix
S
Notice that the displayed eigenmode is not in general piecewise constant. Rather, the
eigenvector is more like vibrational mode of a non-uniform membrane (in fact, they can
be modeled in precisely that way). Also, for all but the stationary distribution, there is a
significant net flow between neighbours, especially in regions where the magnitude of the
spatial gradient of the eigenmode is larger.
4 Perturbation Analysis of EigenFlows
As discussed in the introduction, we seek to identify bottlenecks in the eigenflows associated with long halflives. This notion of identifying bottlenecks is similar to the well-known
max-flow, min-cut theorem. In particular, for a graph whose edge weights represent maximum flow capacities between pairs of vertices, instead of the current conditional transition
probabilities, the bottleneck edges can be identified as precisely those edges across which
the maximum flow is equal to their maximum capacity. However, in the Markov framework, the flow of probability across an edge is only maximal in the extreme cases for
which the initial probability of being at one of the edge?s endpoints is equal to one, and
zero at the other endpoint. Thus the max-flow criterion is not directly applicable here.
Instead, we show that the desired bottleneck edges can be conveniently identified by considering the sensitivity of the flow?s halflife to perturbations of the edge weights (see Fig.
2c). Intuitively, this sensitivity arises because the flow across a bottleneck will have fewer
alternative routes to take and therefore will be particularly sensitive to changes in the edge
weights within the bottleneck. In comparison, the flow between two vertices in a strongly
coupled cluster will have many alternative routes and therefore will not be particularly
sensitive on the precise weight of any single edge.
In order to pick out larger halflives, we will use one parameter, , which is a rough estimate
H
of the smallest halflife that one wishes to consider. Since we are interested in perturbations
which significantly change the current halflife of a mode, we choose to use a logarithmic
scale
in halflife. A simple choice for a function which combines these two effects is : =
: = , where the halflife of the current eigenmode.
H
N
^ of Z , with eigenvalue . This eigenvector decays with
Suppose we have aneigenvector
:XL=
:D = . Consider the effect on d: = of perturbing the affinity
a halflife of F
for the : J = -edge, to . In particular, we show in the Appendix that the
derivative of :f: =M= with respect to , evaluated at
, satisfies
&
:
H=
&
: ^ M ^ #=
^
&
:XL=
: I= :
=
N
: J= elements of eigenvector ^
^ P
&
:
I=
^ P
&
^ P
&
(5)
are the
and :X&&#= are degrees of nodes
(Eq.1). In Figure 2, for a given eigenvector and its flow, we plot the maximum of
absolute sensitivities of all the weights on edges connected to a pixel (not including itself).
Note that the sensitivities are large in the bottlenecks at the border of the foreground and
background.
Here
: J =
5
E IGEN C UTS : A Basic Clustering Algorithm
We select a simple clustering algorithm to test our proposal of using the derivative of the
eigenmode?s halflife for identifying bottleneck edges. Given a value of , which is roughly
H
the minimum halflife to consider for any eigenmode, we iterate the following:
1. Form the symmetric affinity matrix , and initialize .
2. Set !#" $&%()+' *-, ." ) , and set a scale factor / to be the median of
,?>+@
,?>+@
Form the symmetric matrix ; =<
.
=<
3. Compute eigenvectors ACD B , 4D B @ 49E9E9E4FD B
of ;H , with eigenvalues I J ,
'FG
DN
B
for 1$325476869694: .
0#"
IFKLI J
@
IM69696KLI J
'
NQP3R
I
.
4. For
each eigenvector
of ; with halflife O
OTS , compute the halflife sensitivities,
U N
) WVX Y:Z\[^]`_\ab]
Mc for each edge in the graph. Here we use R h2\i\j .
#"
V+dFe.f g
5. Do non-maximal
suppression within each of the computed sensitivities.
That is, suppress
U N
U N
U N
the sensitivity ." ) if there is a strictly more negative value k " ) or ." for some vertex
l k
in the neighbourhood of l ) , or some l in the neighbourhood of l . '
N
6. Compute the sum m
of
t
t
i`/ . We use
hnvu6w2 .
n
U
N
#"
)8o )
#"
'
over all non-suppressed edges
for which
p.1:4Cqr
7. Select the eigenmode D B N8x for which m N8x is maximal.
U N8x
8. Cut all edges p.1:4Cqr in y (i.e. set their affinities to 0) for which #" )0s
this sensitivity was not suppressed during non-maximal suppression.
t
i\/
U
N
."
)s
and for which
9. If any new edges have been cut, go to 2. Otherwise stop.
Here steps (z are as described previously, other than computing the scaling constant { ,
which is used in step to provide a scale invariant threshold on the computed sensitivities.
In step 4 we only consider eigenmodes with halflives larger than |D , with | _ } because
H
this typically eliminates the need to compute the sensitivities for many modes with tiny
values of and, because of the term in : = , it is very rare for eigenvectors with
S
H
halflives smaller than | to produce any sensitivity less than ~ .
H
In step 5 we perform a non-maximal suppression on the sensitivities for the ?C? eigenvector.
We have observed that at strong borders the computed sensitivities can be less than ~ in a
band along the border few pixels thick. This non-maximal suppression allows us to thin this
region. Otherwise, many small isolated fragments can be produced in the neighbourhood
of such strong borders.
In step 6 we wish to select one particular eigenmode to base the edge cutting on at this
iteration. The reason for not considering all the modes simultaneously is that we have
found the locations of the cuts can vary by a few pixels for different modes. If nearby
edges are cut as a result of different eigenmodes, then small isolated fragments can result
in the final clustering. Therefore we wish to select just one eigenmode to base cuts on each
iteration. The particular eigenmode selected can, of course, vary from one iteration to the
next.
The selection strategy in step 6 above
picks out the mode which produces the largest
linearized
: S
H = . That is, we compute
increment in d: S =
S
_
, where
is the change of affinities for any edge left to
-
e#f g
otherwise. Other techniques for selecting a particular mode were
be cut, and
also tried, and they all produced similar results.
This iterative cutting process must eventually terminate since, except for the last iteration,
edges are cut each iteration and any cut edges are never uncut. When the process does
terminate, the selected succession of cuts provides a modified affinity matrix 8 which
has well separated clusters. For the final clustering result, we can use either a connected
components algorithm or the -means algorithm of [5] with set to the number of modes
having large halflives.
6 Experiments
We compare the quality of E IGEN C UTS with two other methods: a -means based spectral
clustering algorithm of [5] and an efficient segmentation algorithm proposed in [1] based on
a pairwise region comparison function. Our strategy was to select thresholds that are likely
to generate a small number of stable partitions. We then varied these thresholds to test the
quality of partitions. To allow for comparison with -means, we needed to determine the
number of clusters a priori. We therefore set to be the same as the number of clusters
that E IGEN C UTS generated. The cluster centers were initialized to be as orthogonal as
possible [5].
The first two rows in Fig. 3 show results using E IGEN C UTS. A crucial observation with
E IGEN C UTS is that, although the number of clusters changed slightly with a change in
H , the regions they defined were qualitatively preserved across the thresholds and corresponded to a naive observer?s intuitive segmentation of the image. Notice in the random
images the occluder is found as a cluster clearly separated from the background. The performance on the eye images is also interesting in that the largely uniform regions around
the center of the eye remain as part of one cluster.
In comparison, both the -means algorithm and the image segmentation algorithm of [1]
(rows 3-6 in Fig. 3) show a tendency to divide uniform regions and give partitions that are
neither stable nor intuitive, despite multiple restarts.
7 Discussion
We have demonstrated that the common piecewise constant approximation to eigenvectors arising in spectral clustering problems limits the applicability of previous methods to
situations in which the clusters are only relatively weakly coupled. We have proposed a
new edge cutting criterion which avoids this piecewise constant approximation. Bottleneck
edges between distinct clusters are identified through the observed sensitivity of an eigenflow?s halflife on changes in the edges? affinity weights. The basic algorithm we propose
is computationally demanding in that the eigenvectors of the Markov matrix must be recomputed after each iteration of edge cutting. However, the point of this algorithm is to
simply demonstrate the partitioning that can be achieved through the computation of the
sensitivity of eigenflow halflives to changes in edge weights. More efficient updates of the
eigenvalue computation, taking advantage of low-rank changes in the matrix Z
from one
iteration to the next, or a multi-scale technique, are important areas for further study.
(a)
(b)
(c)
(d)
(e)
Figure 3: Each column refers to a different image in the dataset shown in Fig. 1. Pairs
of rows correspond to results from applying: E IGEN C UTS with L +|
$+~
and H
(Rows 1&2), -Means spectral clustering where , the number
of clusters, is determined by the results of E IGEN C UTS (Rows 3&4) and Falsenszwalb &
(Rows 5&6).
Huttenlocher ~
Acknowledgements
We have benefited from discussions with Sven Dickinson, Sam Roweis, Sageev Oore and Francisco
Estrada.
References
[1] P. Felzenszalb and D. Huttenlocher Efficiently Computing a Good Segmentation Internation
Journal on Computer Vision, 1999.
[2] R. Kannan, S. Vempala and A. Vetta On clusterings?good, bad and spectral. Proc. 41st Annual
Symposium on Foundations of Computer Science , 2000.
[3] J. R. Kemeny and J. L. Snell Finite Markov Chains. Van Nostrand, New York, 1960.
[4] M. Meila and J. Shi A random walks view of spectral segmentation. Proc. International Workshop on AI and Statistics , 2001.
[5] A. Ng, M. Jordan and Y. Weiss On Spectral Clustering: analysis and an algorithm NIPS, 2001.
[6] A. Ng, A. Zheng, and M. Jordan Stable algorithms for link analysis. Proc. 24th Intl. ACM SIGIR
Conference, 2001.
[7] A. Ng, A. Zheng, and M. Jordan Link analysis, eigenvectors and stability. Proc. 17th Intl. IJCAI,
2001.
[8] P. Perona and W. Freeman A factorization approach to grouping. European Conference on
Computer Vision, 1998.
[9] A. Pothen Graph partitioning algorithms with applications to scientific computing. Parallel
Numerical Algorithms, D. E. Keyes et al (eds.), Kluwer Academic Press, 1996.
[10] G. L. Scott and H. C. Longuet-Higgins Feature grouping by relocalization of eigenvectors of
the proximity matrix. Proc. British Machine Vision Conference, pg. 103-108, 1990.
[11] J. Shi and J. Malik. Normalized cuts and image segmentation. IEEE Transaction on Pattern
Analysis and Machine Intelligence , 2000.
[12] N. Tishby and N. Slonim Data clustering by Markovian Relaxation and the Information Bottleneck Method. NIPS, v 13, MIT Press, 2001.
[13] Y. Weiss Segmentation using eigenvectors: a unifying view. International Conference on
Computer Vision, 1999.
Appendix
We compute the derivative of the log of half-life O of an eigenvalue J with respect to an element o )
of the affinity matrix . Half-life is defined as the power to which J must be raised to reduce the
eigenvalue to half, i.e., J ] 28i . What we are interested is in seeing significant changes in those
half-lives O which are relatively large compared to some minimum half-life O S . So eigenvectors with
half-lives smaller than O S are effectively ignored. It is easy to show that,
p
r
;
p O
ObS9r
J
J
o )
o ) 4 and o ) DB
o ) D B 6
(6)
J
pCJbr
pCJ ]
i5r
Let D B be the corresponding
eigenvector
such that ; D B &JD B , where ; is the modified affinity matrix
,?>M@
,?>M@
(Sec 2). As ; ( <
<
, we can write for all 1 q :
,?>M@
,?>M@
,?>+@
,?>M@
;
)
)
o ) ( <
n <
n
<
4
(7)
<
)
is a matrix of all zeros except for a value of 2 at location pb4br ; p 4 r are degrees of
where
V!e #"@ %$&
the nodes 1 and q (stacked as elements on the diagonal matrix see Sec 2); and
) )
V g #"@ %$&
having non-zero entries only on the diagonal. Simplifying the expression further, we get
,?>+@
,?>M@
,?>+@
,?>M@
,?>+@
;
D
D
D
)
)
o ) DB DB
<
B
B n&B
<
<
<
DB
(8)
,?>M@
,?>+@
,?>+@
D
n&B
<
<
7
DB 6
,?>+@
,?>M@
,?>M@
D
D
D
y <
B
; B
JB , and
Using the fact that <
diagonal, the above equation reduces to:
,?>+@
,?>M@
;
D
D
D
)
)
o ) DB DB
<
B
B n( JB
' <
,?>+@
,?>M@
D
D
D
)
)
<
B
<
B n
JB
The scalar form of this expression is used in Eq.5.
,?>+@
,?>M@
<
,
D
as both
and
are
B
)<
,
) )
D
B 6
(9)
| 2183 |@word stronger:1 grey:1 seek:1 linearized:1 tried:1 decomposition:1 simplifying:1 pg:1 brightness:1 pick:2 initial:6 fragment:3 selecting:1 bc:1 current:5 must:5 numerical:1 partition:3 remove:1 plot:2 update:1 stationary:5 half:10 generative:1 fewer:1 selected:2 intelligence:1 filtered:1 provides:4 node:2 toronto:2 location:2 along:5 symposium:1 combine:1 pairwise:1 inter:1 allan:1 roughly:3 nor:1 multi:1 occluder:2 freeman:1 decreasing:1 automatically:1 considering:4 begin:1 notation:2 bounded:1 moreover:1 mass:3 what:1 eigenvector:13 partitioning:4 unit:1 limit:1 slonim:1 despite:1 uncut:1 might:1 factorization:1 practical:1 implement:1 area:1 significantly:2 convenient:3 refers:1 seeing:1 get:1 selection:1 applying:1 demonstrated:1 center:2 shi:2 go:1 starting:1 sigir:1 identifying:2 insight:1 higgins:1 datapoints:1 stability:1 notion:1 variation:2 increment:1 feel:1 suppose:2 neighbouring:2 dickinson:1 element:5 jk:1 particularly:2 continues:1 cut:14 huttenlocher:2 observed:2 capture:1 region:13 ensures:2 connected:6 acd:1 ideally:1 weakly:4 basis:1 textured:1 various:1 represented:2 stacked:1 separated:3 distinct:1 recomputes:1 sven:1 corresponded:1 whose:1 larger:3 solve:1 otherwise:3 statistic:1 gi:2 itself:2 final:2 advantage:2 eigenvalue:13 net:3 propose:3 maximal:6 rapidly:1 roweis:1 intuitive:2 qr:1 ijcai:1 cluster:28 intl:2 produce:3 derive:1 coupling:1 propagating:1 eq:3 strong:2 c:1 eigenflows:11 indicate:1 direction:2 thick:1 behaviour:1 ots:1 snell:1 strictly:1 proximity:1 around:1 deciding:1 vary:2 smallest:1 purpose:1 proc:5 applicable:1 sensitive:2 largest:1 pcj:1 weighted:1 rough:1 clearly:1 mit:1 gaussian:1 modified:3 rather:1 vibrational:1 rank:1 contrast:1 suppression:4 typically:1 initially:1 perona:1 interested:2 pixel:13 overall:1 priori:2 spatial:1 raised:1 initialize:1 equal:2 construct:1 never:1 having:5 ng:3 represents:1 thin:1 foreground:5 piecewise:8 few:3 neighbour:2 simultaneously:1 tightly:1 attempt:1 ab:1 fd:1 investigate:1 zheng:2 extreme:1 chain:11 edge:48 orthogonal:2 divide:1 walk:5 desired:1 initialized:1 isolated:2 column:1 markovian:3 measuring:1 applicability:1 vertex:19 deviation:1 rare:1 entry:1 uniform:3 front:1 pothen:1 tishby:1 st:1 international:2 sensitivity:20 choose:1 slowly:1 f5:1 corner:1 derivative:3 leading:4 sec:2 coefficient:1 caused:1 break:1 view:3 observer:1 analyze:1 decaying:1 parallel:1 formed:2 largely:1 efficiently:1 ensemble:2 correspond:2 identify:7 succession:1 mgi:2 iterated:1 critically:1 produced:2 mc:1 m5s:1 ed:1 eigenmode:10 definition:1 associated:5 stop:1 dataset:3 recall:1 ut:8 segmentation:11 chakra:2 restarts:1 wei:2 formulation:2 evaluated:1 strongly:1 furthermore:1 just:1 until:1 defines:2 mode:8 eigenmodes:2 quality:2 gray:5 scientific:1 effect:2 normalized:1 hence:1 symmetric:9 white:1 during:1 criterion:2 demonstrate:1 image:15 novel:1 ef:1 common:1 perturbing:1 endpoint:2 b4:1 discussed:1 kluwer:1 interpret:1 refer:1 significant:5 ai:1 meila:1 similarly:1 particle:7 stable:6 base:2 chennubhotla:1 recent:1 perspective:2 reverse:2 route:2 nostrand:1 wv:1 success:1 life:8 minimum:2 estrada:1 determine:2 multiple:1 reduces:1 smooth:1 academic:1 characterized:3 long:4 oore:1 sageev:1 basic:2 vision:4 iteration:9 represent:1 achieved:1 proposal:1 background:6 cropped:1 preserved:1 median:2 crucial:1 w2:1 eliminates:1 unlike:1 undirected:2 db:6 flow:22 jordan:3 ideal:1 easy:1 iterate:1 identified:6 perfectly:1 restrict:1 reduce:2 idea:1 opposite:2 br:1 bottleneck:17 expression:2 york:1 ignored:1 relocalization:1 eigenvectors:21 dark:1 band:1 generate:1 specifies:1 notice:4 sign:1 arising:1 discrete:1 write:1 express:1 key:1 recomputed:1 threshold:4 neither:1 graph:17 relaxation:3 sum:1 appendix:2 scaling:1 annual:1 infinity:1 precisely:2 nearby:1 min:1 vempala:1 relatively:2 department:1 membrane:1 describes:1 across:5 smaller:2 suppressed:2 slightly:1 partitioned:1 remain:1 sam:1 making:2 intuitively:1 invariant:1 computationally:2 equation:1 previously:1 eventually:1 needed:2 tractable:1 keyes:1 observe:1 randomn:1 spectral:12 appropriate:1 neighbourhood:4 alternative:2 clustering:16 cf:1 unifying:1 dfe:1 especially:1 malik:1 added:1 strategy:2 diagonal:5 kemeny:1 affinity:19 gradient:1 link:2 capacity:2 lthe:1 reason:1 kannan:1 assuming:1 length:1 modeled:1 illustration:1 cq:2 equivalently:1 negative:2 suppress:1 motivates:1 perform:2 observation:1 markov:16 finite:1 displayed:1 situation:1 variability:1 precise:1 perturbation:7 varied:1 canada:1 pair:3 namely:2 nip:2 pattern:1 scott:1 including:2 max:2 power:1 demanding:1 representing:1 eye:3 igen:8 coupled:6 ues:1 naive:1 acknowledgement:1 interesting:1 proportional:1 vg:1 h2:1 foundation:1 degree:2 tiny:1 row:6 course:1 changed:1 last:1 bias:2 side:1 allow:1 taking:3 absolute:5 fg:1 van:1 dip:1 boundary:2 transition:10 ending:1 avoids:1 qualitatively:1 transaction:1 cutting:4 confirm:1 assumed:1 francisco:1 spectrum:1 iterative:3 terminate:2 longuet:1 expansion:1 european:1 z_:1 diag:4 jepson:2 border:4 noise:1 eigenflow:9 fig:5 representative:1 benefited:1 fashion:1 wish:4 down:1 theorem:1 british:1 bad:1 decay:7 normalizing:1 grouping:2 workshop:1 effectively:1 magnitude:1 logarithmic:1 simply:3 likely:1 conveniently:1 scalar:1 corresponds:1 chance:1 satisfies:1 acm:1 conditional:2 vetta:1 sorted:1 kli:1 internation:1 change:8 infinite:1 typical:1 except:2 determined:1 tendency:1 select:7 latter:1 arises:1 h5:1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.